ICLR 2026 - Reviews

SubmissionsReviews

Reviews

Summary Statistics

EditLens Prediction Count Avg Rating Avg Confidence Avg Length (chars)
Fully AI-generated 1 (25%) 2.00 3.00 2281
Heavily AI-edited 0 (0%) N/A N/A N/A
Moderately AI-edited 0 (0%) N/A N/A N/A
Lightly AI-edited 1 (25%) 2.00 4.00 3941
Fully human-written 2 (50%) 4.00 4.00 1726
Total 4 (100%) 3.00 3.75 2418
Title Ratings Review Text EditLens Prediction
TRANSPORT-BASED MEAN FLOWS FOR GENERATIVE MODELING Soundness: 2: fair Presentation: 3: good Contribution: 2: fair Rating: 2: reject Confidence: 4: You are confident in your assessment, but not absolutely certain. It is unlikely, but not impossible, that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. This work proposes integrating the Minibatch OT coupling into Mean Flow training. To ensure the training efficiency, various OT computation methods are adopted, including Sinkhorn, Linear OT, and etc. The proposed method is evaluated on image generation (MNIST), shape-conditioned reconstruction (ShapeNet Chairs and ModelNet10), and image-to-image translation (FFHQ), demonstrating improvements over Mean Flow across these tasks. - Extending Mean Flow with Minibatch OT coupling seems to be an interesting approach to improve one-step generative models by leveraging the demonstrated effectiveness of Minibatch OT. - Various experiments have shown that the introduction of OT coupling can improve the generation quality. - The writing and presentation of the manuscript are overall clear and easy to follow. However, with regard to the current manuscript, I still have several major concerns: Limited Novelty: This work primarily combines Mean Flow (Geng et al. 2025) with Minibatch OT (Pooladian et al., 2023; Tong et al., 2023a). Since minibatch OT approaches have been thoroughly explored and proven effective, it is unsurprising that they improve generation quality over Mean Flow. The contribution to the research community appears limited. That said, a thorough evaluation justifying the necessity of Minibatch OT for Mean Flow could still be a valuable contribution. However, I have several questions about the current evaluation. Evaluation Tasks: The evaluation tasks do not follow standard protocols. The original Mean Flow paper uses ImageNet 256×256, a standard benchmark for image generation. This work evaluates only on MNIST for image generation, making it unclear whether the conclusions hold at scale. For 3D shape generation, the choice of a reconstruction task (using the input shape as conditioning) is unclear when unconditional generation could be evaluated instead, following the protocol in LION [A]. Evaluation Baselines: The manuscript primarily compares against Mean Flow alone. However, various works have explored techniques for accelerating inference, including adversarial distillation, consistency models, and shortcut models. The current evaluation does not include comparisons with these baselines. Overall, I recommend rejection due to limited novelty and insufficient evaluation. The work primarily combines existing techniques (Mean Flow and Minibatch OT) in a straightforward manner, yielding expected rather than surprising improvements. The experimental evaluation uses non-standard benchmarks (MNIST instead of ImageNet 256×256, reconstruction tasks instead of unconditional generation) and lacks comparisons with relevant baselines such as adversarial distillation, consistency models, and shortcut models. A substantially revised submission with comprehensive benchmarking on standard tasks and comparisons against the broader landscape of one-step generation methods would be necessary to assess the true contribution of this work. Based on these, I have the following questions: - The evaluation for image generation is limited to MNIST, while the original Mean Flow paper demonstrates results on ImageNet 256×256, which is a standard benchmark in the field. Can you provide experiments on ImageNet 256×256 to demonstrate that OT-MF's improvements generalize to high-resolution, complex image generation tasks? Without this evaluation, it remains unclear whether the benefits of incorporating OT coupling hold at scale. - The manuscript primarily compares OT-MF against Mean Flow. However, there are several other approaches for one-step or few-step generation, including adversarial distillation methods, consistency models, and shortcut models. Can you include comparisons with these baselines to better contextualize OT-MF's performance relative to the broader landscape of accelerated generative modeling techniques? This would help clarify the practical advantages of your approach. Lightly AI-edited
TRANSPORT-BASED MEAN FLOWS FOR GENERATIVE MODELING Soundness: 3: good Presentation: 3: good Contribution: 2: fair Rating: 4: marginally below the acceptance threshold Confidence: 4: You are confident in your assessment, but not absolutely certain. It is unlikely, but not impossible, that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. The paper proposes to improve the training of Mean Flow models by incorporating optimal transport. Specifically, minibatch OT methods (i.e. computing OT on the empirical distribution represented by the minibatch) define a coupling on the elements of a batch that is used to train the models. Intuitively, this results in straighter trajectories that should be simpler to learn with Mean Flows. Experiments on toy data, MNIST, shapes, and paired image-to-image translation are presented. - Clear presentation of mathematics and motivation - Good overview of relation to previous work - Proposes a natural - yet novel - idea. - A diverse set of experiments is performed and some improvements are shown - While a natural idea, the method has limited novelty. - The experiments are limited: MNIST is still a toy data and it is only performed on a latent space of size 4x4x4. All improvements are paid for by significant longer training times per epoch. Therefore, it is unclear how much the use of the OT coupling is really useful here. Typos/comments: - Abstract, first word: Flow-matching -> Flow matching - Page 3: “and introduces stochasticity into the model”. It does not introduce further stochasticity and I would remove that comment. Missing references: - Introduction: Mean Flow is equivalent to Flow Map Matching [1] and they should be cited together [1] https://arxiv.org/abs/2406.07507 Fully human-written
TRANSPORT-BASED MEAN FLOWS FOR GENERATIVE MODELING Soundness: 4: excellent Presentation: 4: excellent Contribution: 1: poor Rating: 2: reject Confidence: 3: You are fairly confident in your assessment. It is possible that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. Math/other details were not carefully checked. This paper proposes optimal transport MeanFlow (OT-MF), which integrates OT-based sampling with the MeanFlow framework for one-step generative modeling. The training samples obtained via OT couplings (e.g., Sinkhorn or Linear OT) exhibit geometrically straighter trajectories compared to independent Gaussian-data pairs. This coupling strategy, when combined with the time-averaged MeanFlow formulation, can reduce inference steps while preserving the fidelity of the multi-step flow process. Experiments on diverse modalities (image generation, image-to-image translation, and point cloud generation) demonstrate that proposed method shows improved sample quality, faster inference, and better trajectory alignment compared to vanilla MeanFlow. - The proposed combination of OT and MeanFlow is well-motivated to improve trajectory straightness and sampling efficiency. - The paper provides comprehensive background and related work on Flow Matching, OT-based methods, and MeanFlow. - Experiments are conducted on diverse tasks and demonstrate consistent improvements in sample quality and efficiency. - Incorporation of scalable OT solvers enhances computational overhead without major performance loss. - Novelty is weak. The main idea (combining existing OT-based coupling with MeanFlow) is conceptually straightforward. While effective, it primarily extends prior techniques rather than introducing a fundamentally new theoretical framework. - The proposed method still fails to accurately capture the data distribution, even in a simple 2D toy dataset (Figure 2). - (Minor comment) The qualitative difference between MF and OT-MF is subtle in Figure 1 illustration. Using multiple (x0, x1) trajectories or average path visualizations could better highlight the ‘trajectory straightening’ effect. - The proposed method primarily combines two existing ideas, OT-based flow matching and MeanFlow. Beyond this integration, is there any novel algorithms, theoretical contribution, or architectural enhancement introduced in the paper? - In Table 1, why does Sinkhorn OT often outperform the proposed OT-MF in terms of Wasserstein-2 distance? - Is this method still effective in higher-dimensional or sequential generative tasks, such as audio synthesis or text-to-video generation? Fully AI-generated
TRANSPORT-BASED MEAN FLOWS FOR GENERATIVE MODELING Soundness: 2: fair Presentation: 3: good Contribution: 2: fair Rating: 4: marginally below the acceptance threshold Confidence: 4: You are confident in your assessment, but not absolutely certain. It is unlikely, but not impossible, that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. The paper integrates optimal transport and mean flows for better one and few steps generation with geometry-aware learning. The author claims the unified framework performs efficiently on few different downstream tasks such as toy example, image generation, point cloud generation and image-to-image translation. 1.The paper is well written and easy to follow. 2.The paper leverages OT coupling mechanisms and mean flows on the training objective to enforce globally optimal source–target alignments, leading to straighter and more stable transport trajectories. 3.The work achieved good quality performance in multiple tasks with few step generation. 1.Limited novelty and incremental improvement: The main contribution lies in combining established OT techniques with mean flows. The paper lacks new theoretical developments or formal analysis to differentiate it from prior OT-regularized flow-matching approaches. The authors are encouraged to clarify what is theoretically or algorithmically new beyond the combination itself. 2.Incomplete baselines of one/few steps generation: The authors achieved better performance compared to meanflow and other OT techniques but the lack of comparison with recent one/few steps SOTA generative models limits the impact of the proposed framework. Models such as consistency/distillation models or few step-diffusion or flow matching baselines are not cited or discussed [1,2,3]. 3. Lack of hyperparameter analysis: The paper does not discuss how the hyperparameters (e.g. batch size, learning rate and etc) are selected. The paper should explain the procedure or tuning strategy of the models prior to the experiments. References: [1] Song, Yang, et al. "Consistency models." (2023). [2] Meng, Chenlin, et al. "On distillation of guided diffusion models." Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2023. [3] Sauer, Axel, et al. "Adversarial diffusion distillation." European Conference on Computer Vision. Cham: Springer Nature Switzerland, 2024. note the weakness above Fully human-written
PreviousPage 1 of 1 (4 total rows)Next