ICLR 2026 - Reviews

SubmissionsReviews

Reviews

Summary Statistics

EditLens Prediction Count Avg Rating Avg Confidence Avg Length (chars)
Fully AI-generated 1 (33%) 0.00 5.00 2756
Heavily AI-edited 1 (33%) 0.00 1.00 139
Moderately AI-edited 0 (0%) N/A N/A N/A
Lightly AI-edited 1 (33%) 0.00 5.00 89
Fully human-written 0 (0%) N/A N/A N/A
Total 3 (100%) 0.00 3.67 995
Title Ratings Review Text EditLens Prediction
A Diffusion-Based Data Augmentation Approach for Synthetic Human Portraits Dataset Soundness: 1: poor Presentation: 1: poor Contribution: 1: poor Rating: 0: Confidence: 1: You are unable to assess this paper and have alerted the ACs to seek an opinion from different reviewers. The paper appears to be incomplete The paper appears to be incomplete The paper appears to be incomplete The paper appears to be incomplete Heavily AI-edited
A Diffusion-Based Data Augmentation Approach for Synthetic Human Portraits Dataset Soundness: 1: poor Presentation: 1: poor Contribution: 1: poor Rating: 0: Confidence: 5: You are absolutely certain about your assessment. You are very familiar with the related work and checked the math/other details carefully. This paper is an incomplete submission. N.A. This paper is an incomplete submission. N.A. Lightly AI-edited
A Diffusion-Based Data Augmentation Approach for Synthetic Human Portraits Dataset Soundness: 1: poor Presentation: 1: poor Contribution: 1: poor Rating: 0: Confidence: 5: You are absolutely certain about your assessment. You are very familiar with the related work and checked the math/other details carefully. The paper proposes a diffusion-based image-to-image augmentation method that attempts to preserve facial/acupoint landmarks. The pipeline uses a combination of Stable Diffusion (SD-1.5), an IP-Adapter, an IC-Light module, and a choice of noise level injection to control how much the original image is retained. The motivation is to generate new visual diversity (background, lighting, skin tone, hair) while keeping acupoint labels valid, so that one can reuse existing labels without relabeling. The authors build a new “AcuSim” dataset (225 subjects × 44 views ≈ 9,900 images) and report very high accuracy (~99 %) on a CNN for acupoint classification and small landmark drift (5–8 pixels for most points). 1) The goal is compelling: reduce labeling burden by generating new but label-consistent data. 2) The idea of mixing partial guidance (injection of original image at noise level) to maintain structure is promising. 1) The paper is not ready for submission for instance in pp 5 it mentioned that "Figure 2: Enter Caption". As we can see in pp 2 in "Figure 1: General Workflow" 2) The method is largely an engineering combination of existing modules (Stable Diffusion, IP-Adapter, lighting control). The novelty lies in their orchestration for this task, but that might be considered incremental. 3) The “99 % accuracy” result is too good to be believable without more context: which baseline, which dataset split, how many classes, variance across runs, and is the model overfitting? 4) There is no test of generalization: i.e. train on augmented data, test on new subjects or different lighting/pose settings. Does the augmentation help or hurt out-of-distribution generalization? Missing relevant references in literature review 1) Context-guided Responsible Data Augmentation with Diffusion Models 2) Effective Data Augmentation With Diffusion Models 3) Diffusion models: A comprehensive survey of methods and applications 4) GenMix: Effective Data Augmentation with Generative Diffusion Model Image Editing 5) DiffuseMix: Label-Preserving Data Augmentation with Diffusion Models 1) Compare the proposed method vs. classical augmentation (flip, color jitter, random crop) and, ideally, a GAN-based or identity-preserving diffusion baseline on the acupoint task. 2) Provide ablation experiments: how much each module (IP-Adapter, IC-Light, injection strength) contributes, and report performance vs drift tradeoffs across settings. 3) Test generalization: train with augmented + original, evaluate on unseen subjects, unseen lighting, or even a held-out real dataset to ensure no overfitting. 4) If possible, show that downstream tasks beyond acupoint classification (e.g. landmark regression, localization) benefit from the augmentation. Fully AI-generated
PreviousPage 1 of 1 (3 total rows)Next