ICLR 2026 - Reviews

SubmissionsReviews

Reviews

Summary Statistics

EditLens Prediction Count Avg Rating Avg Confidence Avg Length (chars)
Fully AI-generated 0 (0%) N/A N/A N/A
Heavily AI-edited 1 (25%) 2.00 3.00 1617
Moderately AI-edited 0 (0%) N/A N/A N/A
Lightly AI-edited 1 (25%) 2.00 4.00 3901
Fully human-written 2 (50%) 1.00 3.50 4884
Total 4 (100%) 1.50 3.50 3821
Title Ratings Review Text EditLens Prediction
Adaptive Drug-Drug Interaction Prediction via Gauge-Aware Graph Representation and Distribution Alignment Soundness: 2: fair Presentation: 1: poor Contribution: 1: poor Rating: 2: reject Confidence: 4: You are confident in your assessment, but not absolutely certain. It is unlikely, but not impossible, that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. This paper proposed GraphPharmNet for the drug-drug interaction (DDI) prediction with limited data and unpredictable distribution shift. The core of GraphPharmNet is to link a compact gauge-aware graph encoder and light-weight distribution alignment objectives. The proposed GraphPharmNet method was compared with six state-of-the-art GNN-based methods on the DrugBank dataset. S1: The proposed GraphPharmNet is to combine gauge-aware graph encoder and light-weight distribution alignment for the drug-drug interaction (DDI) prediction under the conditions of limited data and distribution shift. S2: In reported experiments, the proposed GraphPharmNet outperforms existing GNN-based DDI prediction methods in terms of accuracy, micro-F1, and macro-F1. W1: The abstract is too obscure and difficult to understand. Too many irrelevant statements affect readability and make it difficult to understand the core ideas. W2: The core of the work is leakage-safe gauge-aware message passing with alignment. However, too many concepts and theorems are discussed but lack a reasonable explanation of the motivation for proposing them. W3: The paper could be strengthened from more advanced DDI prediction methods such as DMFDDI and further validation along different DDI datasets. - How does the proposed gauge-aware graph encoder based on edge orthogonal transport and kernelized neighborhood weighting address the data scarcity problem in DDI prediction? How does lightweight distribution alignment handle unpredictable distribution shifts in DDI prediction? - What does MMD stand for? This abbreviation should be clearly defined before use. - The description in Figure 1 does not match the content of Figure 1. For example, the middle section: a gauge-aware encoder transports neighbor features via orthogonal transforms Ruv and weights contributions by Kuv before aggregation. However, the middle section of Figure 1 focuses on DDI subgraph generation and knowledge subgraph generation, rather than the so-called gauge-aware graph encoder. - The core of the work is leakage-safe gauge-aware message passing with alignment. However, too many concepts and theorems are discussed but lack a reasonable explanation of the motivation for proposing them. - The paper could be strengthened from more advanced DDI prediction methods such as DMFDDI and further validation along different DDI datasets such as TWOSIDES. - A comprehensive ablation study to dissect the contribution of each component (i.e., Gauge-Awareness, Leakage-Safe Protocol, and Distribution Alignment) is needed. Fully human-written
Adaptive Drug-Drug Interaction Prediction via Gauge-Aware Graph Representation and Distribution Alignment Soundness: 2: fair Presentation: 2: fair Contribution: 2: fair Rating: 2: reject Confidence: 3: You are fairly confident in your assessment. It is possible that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. Math/other details were not carefully checked. This paper proposes GraphPharmNet, a framework for drug-drug interaction (DDI) prediction addressing data scarcity and distribution shifts. Key contributions include: Per-edge orthogonal transports align neighbor features before aggregation, claimed to enhance stability under local basis changes. Uses MMD and entropic optimal transport (OT) on deterministic partitions of training data to handle source heterogeneity. Temporarily removes target edges during training ("drop-edge-by-target") to prevent information leakage. On-the-fly generation of orthogonal transports reduces memory by 60× vs. naive caching. Evaluated on a DrugBank-Hetionet graph, the method reportedly outperforms GNN baselines. The "drop-edge-by-target" protocol is a rigorous solution to edge-label leakage. Emphasis on deterministic splits and label mapping aids reproducibility. "Gauge-awareness" is ambiguously defined—is it a smoothing regularizer or an approximate symmetry? Formal guarantees are missing. Comparisons are limited to generic GNNs, excluding recent DDI-specific models. Only one dataset is tested. Cold-start/relation-biased shift experiments are mentioned but not shown. Distribution alignment is confined to training data, undermining its relevance to real-world shifts. Ablate orthogonal transports: How does performance change on shifted data (e.g., new relation types) without R_uv? Provide metrics for "stability" (e.g., variance in embeddings under perturbations). Why align only training partitions? Compare to aligning training with a synthetic target domain. Show if MMD/OT improves macro-F1 on cold-start tasks. Heavily AI-edited
Adaptive Drug-Drug Interaction Prediction via Gauge-Aware Graph Representation and Distribution Alignment Soundness: 2: fair Presentation: 2: fair Contribution: 2: fair Rating: 2: reject Confidence: 4: You are confident in your assessment, but not absolutely certain. It is unlikely, but not impossible, that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. The authors proposed a framework that combines gauge-aware message passing with training-only distribution alignment, enabling more stable and generalizable representation learning in drug knowledge graphs. The authors proposed a framework that combines gauge-aware message passing with training-only distribution alignment, enabling more stable and generalizable representation learning in drug knowledge graphs. 1. On “a subtle but pervasive type of information leakage.” To my knowledge, most KG-augmented DDI methods (e.g., using DRKG, Hetionet) explicitly remove all test-set DDI edges from the KG prior to training/evaluation. In that case, how does leakage still occur? Could you provide a concrete, reproducible example demonstrating the leakage pathway you have in mind, so readers can verify that leakage persists even after removing test DDI edges? 2. On “Inconsistent Evaluation Practices.” You argue that prior work suffers from inconsistent evaluation; however, many recent papers report five-fold cross-validation, which appears consistent at first glance. Could you clarify what specific inconsistencies you refer to? In addition, for the two points on handling multi-type interactions and directed vs. undirected edges, do you have strong empirical evidence or practical case studies that illustrate clear failures under common setups? Concrete examples would make these claims much more convincing. 3. The font in Figures 2 and 3 is quite small on printouts and standard screens. I recommend enlarging axis labels, legends, and annotations to improve readability. 4. On baseline citations and recency. Could you double-check the citations for the compared baselines? To my understanding, KGNN was introduced at IJCAI 2020, DDKG in Briefings in Bioinformatics 2022, SumGNN in Bioinformatics 2021, and LaGAT in Bioinformatics 2022. In your manuscript these are cited with 2024–2025 references. Are you certain those attributions are correct? Also, given that these baselines are relatively older, it would be helpful to include or discuss more recent (2024–2025) methods to strengthen the comparison. 5. Lack of interpretability of the model 1. On “a subtle but pervasive type of information leakage.” To my knowledge, most KG-augmented DDI methods (e.g., using DRKG, Hetionet) explicitly remove all test-set DDI edges from the KG prior to training/evaluation. In that case, how does leakage still occur? Could you provide a concrete, reproducible example demonstrating the leakage pathway you have in mind, so readers can verify that leakage persists even after removing test DDI edges? 2. On “Inconsistent Evaluation Practices.” You argue that prior work suffers from inconsistent evaluation; however, many recent papers report five-fold cross-validation, which appears consistent at first glance. Could you clarify what specific inconsistencies you refer to? In addition, for the two points on handling multi-type interactions and directed vs. undirected edges, do you have strong empirical evidence or practical case studies that illustrate clear failures under common setups? Concrete examples would make these claims much more convincing. 3. The font in Figures 2 and 3 is quite small on printouts and standard screens. I recommend enlarging axis labels, legends, and annotations to improve readability. 4. On baseline citations and recency. Could you double-check the citations for the compared baselines? To my understanding, KGNN was introduced at IJCAI 2020, DDKG in Briefings in Bioinformatics 2022, SumGNN in Bioinformatics 2021, and LaGAT in Bioinformatics 2022. In your manuscript these are cited with 2024–2025 references. Are you certain those attributions are correct? Also, given that these baselines are relatively older, it would be helpful to include or discuss more recent (2024–2025) methods to strengthen the comparison. 5. Lack of interpretability of the model Lightly AI-edited
Adaptive Drug-Drug Interaction Prediction via Gauge-Aware Graph Representation and Distribution Alignment Soundness: 1: poor Presentation: 1: poor Contribution: 1: poor Rating: 0: Confidence: 3: You are fairly confident in your assessment. It is possible that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. Math/other details were not carefully checked. The paper aims to revisit the DDI prediction problem under data scarcity and distribution shift. It introduces GraphPharmNet, a framework combining a gauge-aware graph encoder with lightweight distribution alignment strategies. The work claims novelty in using per-edge orthogonal transports (computed via a shared edge MLP) to align neighborhood features before aggregation, promoting stability without requiring strict O(d)-equivariance. On the training side, the authors aims to design a leakage-safe protocol with training-only domain alignment (using MMD and optional entropic-OT), ensuring fair evaluation and reproducibility. Experiments on a large merged DrugBank–Hetionet dataset show consistent gains over baseline GNNs (though many SOTA models are excluded from the comparison), though the approach’s scalability and computational trade-offs warrant closer scrutiny. Ovwerall, I think, this paper presents an interesting idea and a commendable effort, but the clarity and coherence of presentation are weak, and the experiments do not sufficiently support the claims. Overall, it lacks sound motivation, depth, and analytical rigor. Strengthening the structure, related work, and experimental validation would greatly improve its suitability for ICLR. 1. Introduces a gauge-aware graph encoder that integrates orthogonal transport mechanisms in message passing, offering a fresh perspective on stabilizing feature aggregation in molecular graphs. 2. Their proposed leakage-safe training protocol and training-only distribution alignment demonstrate careful attention to experimental rigor and reproducibility, which are often overlooked in DDI prediction research. 3. Aims to address practical challenges of distribution shift and data scarcity in biomedical graphs, an underexplored yet important direction for improving real-world drug–drug interaction prediction. 1. The problem doesn't seem well-motivated or some related literature is missing. The discussion contains several problems in current DDI works, but their approch and modeling choices, doesn't seem to have any relation to these. 2. Informaiton in lines 069-097 seems out of context without any references and problem formualtion; thus very hard to understand and relate. 3. Several relevant DDI works are missing from the discussion. DOIs: 10.1145/3511808.3557648, 10.48550/arXiv.1905.00534, 10.1093/bib/bbab441, 10.48550/arXiv.2403.17210, 10.1093/bib/bbab133, 10.1093/bib/bbac597, 10.48550/arXiv.2209.09941, 10.1039/D2SC02023H, 10.48550/arXiv.2508.06576. 4. In Figure 2 and 3, mention the exact values, increase font sizes as it is not mreadable in current format. Tables are preferred. 5. In Section 5.3 : Baselines, it is not clear why you didn't compare it with existing GNN-based DDI models (DOIs: 10.1145/3511808.3557648, 10.48550/arXiv.1905.00534, 10.1093/bib/bbab441, 10.48550/arXiv.2403.17210, 10.1093/bib/bbab133, 10.1093/bib/bbac597, 10.48550/arXiv.2209.09941, 10.1039/D2SC02023H)? Also, connecting to Figure 2, exisitng DDI models is reported to gain 90-98% accuracy, 99% AUROC in 2023. So, a discussion is expected on how this work differencites from them in terms of application and provides new insights to DDI researchers. 6. Results section is very narrow and do not justify a lot of claimed contributions. A lot of theoretical contributions are made, but the result section do not reflect how these things are actually making a real impact in the modeling; or, how these addresses a practical problem in DDI. 6. Ablations are not clear. Also, more detailed ablaitons are expected, concerning the contribution of the paper. For example, study the impact of *gauge-aware*ness, the alignment objective (MMD/OT), and more. A lot of things are introduced, but their impact or necesity was not establsihed proeprly. 7. Contributions state *an efficiency- and memory-conscious implementation with on-the-fly transport generation that scales to graphs with millions of edges without prohibitive memory requirements*, but there is not scaling studies presented. 8. Presentation and writing quality is poor. Very hard to foillow along and understand. Paper structure, word choices and content flow should be improved. 1. The section in lines 044-060 seems disconnected to 060-064. Explain in details both motivators you mentioned, *(i) distribution shift robust against the heterogeneous nature of sources of integrated knowledge and (ii) representation choice safe*. I didn't find any prior discussion to *distribution shift*, what it is in the context of DDI? how and why it happens? and how and why it matters? Also, the second point is unclear too. What do you meant by "*whichfollows a trial-safe policy of experimental practice preventing any unfair performance evaluation.*"? As per current information, I don't see these methodological requirements being established properly. 2. In Line 069, before discussing the technical terms like *edge*, maybe the graph and problem preliminaries should be introduced; so that the readers can understand the situation better. 3. L70/71 says,*the heterogeneous graph integration problem is a fundamental problem* - any related works or references? 4. What does it mean by *we achieve this by **providing an effective stability in stabilizers** when performing local reparameterizations*? Mention the alhgorithm/strategy directly. 5. Lines 75/76 says, *"Throughout, we take **gauge-aware** to mean stabilization under local changes of basis"*. What is *gauge*? Why should we do *gauge-aware* modeling? What is the techincal or application-wise impactg of it? 6. What is MMD in abstract (L026) and L79/80? It should be introduced before using. 7. L80/81 says, *"prevent mis-imbalance of the internal distributions of them often resulting from merging data sources"*. - any related works or references? 8. Add sufficient references in lines 069-097. Also, clarify. Currently, it is very hard to understand. 9. *"It is possible to do exactly equivariant layers with orthogonal basis changes (e.g. with radial nonlinearities and scalar maps)."* - OK, But, why do we need these? 10. What is the size of $d_0$? In L193/194, it says, *"Each drug $v$ has initial features"*. Whjat about $u$? 11. Line 196 says, *for single-label, multi-class DDI type prediction*. In DDI, interaction types ($y$ in this paper) has massive class imbalance (in DrugBank's 86 interaction types) (DOI: 10.48550/arXiv.2508.06576.). Only top 10-15 interaction type has good representation and takes around 70-80% of the data. How do you handle this issue? 12. Improve abslations as stated in weakness. 13. Add missing references and related discussion. Also, why you didn't compare it with existing GNN-based DDI models? Those models seem to outperform this work, so, in practical application, how does this work makes meaningful contribution? I meant, those DDI models are already performing at 98-99% accuracy. Why od we need another model and what problem does it solve that justifies the need of this work? 14. Provide scaling studies to justify contribution iv. 15. In justification of *leakage-safe*, provide relevant references that proves other SOTA DDI models do have this problem. Fully human-written
PreviousPage 1 of 1 (4 total rows)Next