ICLR 2026 - Reviews

SubmissionsReviews

Reviews

Summary Statistics

EditLens Prediction Count Avg Rating Avg Confidence Avg Length (chars)
Fully AI-generated 3 (75%) 4.67 3.67 3033
Heavily AI-edited 0 (0%) N/A N/A N/A
Moderately AI-edited 1 (25%) 6.00 3.00 2220
Lightly AI-edited 0 (0%) N/A N/A N/A
Fully human-written 0 (0%) N/A N/A N/A
Total 4 (100%) 5.00 3.50 2830
Title Ratings Review Text EditLens Prediction
MeInTime: Bridging Age Gap in Identity-Preserving Face Restoration Soundness: 3: good Presentation: 3: good Contribution: 2: fair Rating: 6: marginally above the acceptance threshold Confidence: 4: You are confident in your assessment, but not absolutely certain. It is unlikely, but not impossible, that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. This paper presents MeInTime, a diffusion-based face restoration framework designed to handle cross-age reference-based restoration. Traditional reference-based methods assume age alignment between degraded and reference images, which limits their applicability in real-world scenarios (e.g., historical photos). MeInTime decouples the modeling of identity and age: during training, identity features are learned through reference embeddings and Gated Residual Fusion modules; during inference, an Age-Aware Gradient Guidance strategy is introduced to steer restoration toward the desired age manifold without retraining. Experiments on same-age and cross-age datasets show improved identity preservation and age consistency compared to existing baselines. + The paper extends reference-based face restoration to the cross-age domain, which is an underexplored but practically relevant scenario. + Separating identity learning and age guidance is conceptually elegant and avoids conflicts between identity and age signals. + The Gated Residual Fusion module is well-motivated and effectively stabilizes identity–structure fusion. + The authors benchmark on both same-age and cross-age datasets and introduce age-consistency metrics, providing clear evidence of quantitative gains. + Visual examples demonstrate that the model generates more age-consistent restorations compared to prior works. – Most degraded samples used for visualization and evaluation are synthetically generated with severe distortions (e.g., heavy blur or compression). Under such extreme degradation, facial wrinkles and texture cues are largely lost, making it unreliable to infer age semantics from low-quality inputs. This could easily cause instability or misalignment between estimated and target ages in practice. – The method introduces a gradient-based optimization process during inference to achieve age control. However, it is not clear whether this method is better than those using face editing-based solutions on the reference image or the restored images. It remains unclear why such simpler or more direct methods are not considered or compared. – The AGE metric depends on a pretrained estimator, which might not correlate well with perceptual aging. Additional perceptual studies or human evaluations would strengthen the claims. - The so-called ID-preserving sampling in Eq. 8 effectively modifies the denoising trajectory based on facial feature gradients, which is conceptually similar to facial attribute editing or latent direction control seen in previous works such as [r1,r2] or other latent manipulation methods. The distinction between this “sampling” and conventional attribute-based editing is not articulated. Maybe it is better to add some face editing works in Section 2. [r1] When StyleGAN Meets Stable Diffusion: a w+ Adapter for Personalized Image Generation [r2] LEDITS++: Limitless Image Editing using Text-to-Image Models - Many degraded samples in the paper are heavily distorted. How does the model perform under moderate degradations, where age cues are still partially available? - Why not apply a face editing approach to either (a) edit the reference image to match the target age before restoration, or (b) perform age editing after restoration? Would this achieve comparable or even better age consistency with less computational overhead? - How stable is the Age-Aware Gradient Guidance when the reference and degraded ages differ drastically (e.g., 20s vs. 80s)? Does it sometimes produce over-aging or artifacts? Fully AI-generated
MeInTime: Bridging Age Gap in Identity-Preserving Face Restoration Soundness: 3: good Presentation: 3: good Contribution: 2: fair Rating: 6: marginally above the acceptance threshold Confidence: 3: You are fairly confident in your assessment. It is possible that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. Math/other details were not carefully checked. This paper introduces MeInTime, a novel diffusion-based framework that significantly advances reference-based face restoration by effectively tackling the challenging cross-age scenario. The key innovation lies in its decoupled modeling of identity and age conditions: it injects identity features via a dedicated attention mechanism during training, and at inference, employs a training-free Age-Aware Gradient Guidance to steer the generation towards the target age. Extensive experiments confirm that MeInTime outperforms existing methods, achieving superior identity fidelity and age consistency simultaneously. +Pioneering Cross-Age Reference-Based Framework: This work introduces the first reference-based face restoration framework specifically designed for cross-age scenarios, effectively extending the capability of existing methods from same-age to cross-age restoration by incorporating target age prompts. + Novel Disentangled Training-Inference Strategy: The proposed method employs a decoupled approach that separately handles identity preservation during training through dedicated attention mechanisms, and age consistency during inference via training-free Age-Aware Gradient Guidance, effectively resolving identity-age conflicts. + Gated Residual Fusion modules dynamically integrate structural features from degraded inputs with identity representations in a content-aware manner. + Plug-and-play Age-Aware Gradient Guidance steers generation toward target age semantics without retraining. + Comprehensive experimental validations show superior performance in visual quality, identity preservation, and age consistency compared to existing approaches. - According to Table 1, the performances of the proposed method are not always the best. The authors should explain the reasons in detail. - The authors do not compare the speed and the number of parameters of the proposed method, compared to existing methods. - The authors do not present the failure cases of the proposed method. I think it is better to analyze the limitations. - The fonts in the Figures are too small. - The effectiveness of the Age-Aware Gradient Guidance is not verified in the ablation studies. Please see the weaknesses. Moderately AI-edited
MeInTime: Bridging Age Gap in Identity-Preserving Face Restoration Soundness: 2: fair Presentation: 3: good Contribution: 3: good Rating: 4: marginally below the acceptance threshold Confidence: 4: You are confident in your assessment, but not absolutely certain. It is unlikely, but not impossible, that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. This paper proposes MeInTime, a diffusion-based reference-guided face restoration framework aimed at handling cross-age scenarios where the reference and degraded images of the same person have large age gaps. The key idea is to decouple identity and age conditioning: identity embeddings are injected during training through a Gated Residual Fusion (GRF) module, while age consistency is adjusted during inference using a training-free Age-Aware Gradient Guidance. Solid engineering design, the combination of GRF for stable identity fusion and age-aware gradient guidance is well implemented and empirically effective. Clear structure and writing,the paper is clearly written, visually engaging, and the methodology section is easy to follow. Training-free age control: The gradient-based age guidance is conceptually clean and avoids extra finetuning. 1. The cross-age reference setting is a rare and contrived use case. It’s unclear how many real restoration tasks actually require explicit age matching. The work does not convincingly show that this setting matters beyond a few illustrative examples. The paper starts from the observation that current reference-based face restoration methods assume the reference and degraded faces are of similar age. While this is technically true, the practical importance of bridging “cross-age” gaps in face restoration is quite limited. 2. Experiments rely on synthetically degraded data; no evaluation on truly degraded or historical photos, which weakens the claim of “real-world generalization.” 3. The paper does not report inference time or computational overhead. Given that the proposed Age-Aware Gradient Guidance involves multiple iterative optimization steps, efficiency could be a major concern for practical applications. Quantitative timing or FLOPs comparison is missing. 4. While automatic metrics (FID, MUSIQ, AGE MAE) are presented, subjective evaluations (user or identity verification studies) are missing, which are important for human-centric tasks like face restoration. 5.The related work section overlooks several important prior approaches: [1].Face Super-Resolution Guided by 3D Facial Priors [2].Rethinking Deep Face Restoration 1. Interestingly, Table 1 shows that on cross-age restoration, the reference-free baseline CodeFormer achieves higher PSNR and SSIM than the proposed MeInTime, despite lacking reference information. This suggests that the inclusion of reference images may actually harm reconstruction fidelity when the reference and degraded faces differ significantly in age. The authors should analyze this phenomenon more carefully, as it weakens the central claim that MeInTime “bridges” the age gap effectively. Fully AI-generated
MeInTime: Bridging Age Gap in Identity-Preserving Face Restoration Soundness: 3: good Presentation: 3: good Contribution: 3: good Rating: 4: marginally below the acceptance threshold Confidence: 3: You are fairly confident in your assessment. It is possible that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. Math/other details were not carefully checked. The paper proposes MeInTime, a diffusion-based framework for identity-preserving face restoration that explicitly addresses the challenge of large age gaps between degraded input images and available reference images. MeInTime separates identity and age conditioning: it injects robust identity features during training, while, during inference, it leverages a novel Age-Aware Gradient Guidance mechanism based on textual prompts to control aging, decoupling identity from age in the generative process. 1.The paper targets a practically relevant and previously underexplored problem: high-fidelity, identity-preserving face restoration when only cross-age references are available. 2. The use of a decoupled training-inference strategy—training for identity preservation and introducing age controllability at inference—is a thoughtful response to the lack of large-scale cross-age paired datasets, as discussed with supporting data in Appendix B/Figure 12 (Page 15–16). 1. Limited Theoretical Analysis of Attribute Decoupling: The paper claims that identity and age are decoupled by design (identity during training, age only via gradient guidance at inference), yet it lacks a more principled investigation or proof of whether and to what extent this decoupling is reliably achieved. For example, no formal analysis or visualization is provided to demonstrate that the injected identity embeddings or the age gradients are indeed orthogonal in the learned space. Without explicit investigation, it remains somewhat speculative whether the method fully avoids entanglement between age and identity, especially under distribution shifts. 2. Structural and Optimization Details May Hinder Reproducibility: While implementation details are provided, several critical elements are only broadly sketched. For example, the precise effect of the GRF module hyperparameters, the initialization process for identity/token projection, and the inferred scaling of guidance during inference are not dissected in depth. Additionally, Algorithm 1 might benefit from being more explicit on the stopping criteria, initialization of age/generic prompts, and GRF interaction during inference. 1. Can the authors provide a principled analysis (e.g., mutual information, attention visualization, or orthogonality in feature space) demonstrating that their identity embeddings remain robust (and do not leak age information) when reference images cover very wide age gaps? Empirical or theoretical clarification here would strengthen the claim of identity-age decoupling. 2. What is the actual computational overhead (in wall-clock time or FLOPs) for MeInTime during inference compared to, say, FaceMe or RestorerID, particularly under the Gradient Guidance with multiple optimization passes? Is the method practical for real-time or high-throughput use cases? Fully AI-generated
PreviousPage 1 of 1 (4 total rows)Next