ICLR 2026 - Reviews

SubmissionsReviews

Reviews

Summary Statistics

EditLens Prediction Count Avg Rating Avg Confidence Avg Length (chars)
Fully AI-generated 0 (0%) N/A N/A N/A
Heavily AI-edited 0 (0%) N/A N/A N/A
Moderately AI-edited 0 (0%) N/A N/A N/A
Lightly AI-edited 2 (67%) 5.00 4.00 1984
Fully human-written 1 (33%) 8.00 2.00 1188
Total 3 (100%) 6.00 3.33 1718
Title Ratings Review Text EditLens Prediction
Characteristic Root Analysis and Regularization for Linear Time Series Forecasting Soundness: 3: good Presentation: 4: excellent Contribution: 3: good Rating: 8: accept, good paper Confidence: 2: You are willing to defend your assessment, but it is quite likely that you did not understand the central parts of the submission or that you are unfamiliar with some pieces of related work. Math/other details were not carefully checked. This work presents a comprehensive study of linear models for time series forecasting, focusing on the role of characteristic roots in shaping model expressivity. Experiments validate our theoretical claims and demonstrate the effectiveness of both methods across a range of forecasting tasks. 1. The research motivation of this manuscript is clear and its content is very rich。 2. This manuscript is theoretical. 3. Sufficient experiments have verified the effectiveness of the proposed method. 1. Theoretical analysis is an advantage of this work, but it can easily make it difficult for readers to understand. It is necessary to provide intuitive explanations in the key derivation process 2. The manuscript lacks a comprehensive discussion of related work, which hampers readers, particularly those less familiar with the domain. Thus, it remains unclear how the proposed approach advances the state of the art or distinguishes itself from existing methods. 1. What is the transferability of the method?Can linear models be used in other areas than time series learning? 2. Why is the frequency-domain linear layer necessary for Root-Purge instead of a mere implementation detail? Fully human-written
Characteristic Root Analysis and Regularization for Linear Time Series Forecasting Soundness: 2: fair Presentation: 2: fair Contribution: 3: good Rating: 4: marginally below the acceptance threshold Confidence: 4: You are confident in your assessment, but not absolutely certain. It is unlikely, but not impossible, that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. This paper analyzes linear time series forecasting via characteristic roots , identifying that noise creates "spurious roots" and a "data scaling law" that demands excessive data to mitigate this noise. To improve robustness and data efficiency, it proposes two structural regularization strategies: (1) post-hoc Rank Reduction (RRR) and (2) a novel "Root Purge" method. Root Purge is an adaptive training regularizer that encourages the model to learn a null space for the estimated noise (residuals). 1.Originality and Clarity: The "Root Purge" loss function is a novel, elegant, and intuitive regularization technique. 2.Quality and Significance: The paper provides a valuable insight into why simple linear models fail in noisy, low-data regimes. 1.Hyperparameter Sensitivity: The methods' performance relies heavily on tuning $\rho$ or $\lambda$. This data-specific tuning is difficult to estimate a priori and undermines the claim of robustness. 2.Contingent on Low-Rank Assumption: The methods are motivated by an assumption that the true signal is low-rank. However, on large, complex datasets, the paper's results show the optimal rank is often full , and the proposed methods offer minimal benefit . This limits the methods' generality. 3.Theory-Practice Gap (Trends): The theoretical analysis assumes homogeneous equations (constant/zero bias) . This is a poor fit for real-world data with non-stationary, dynamic trends. 1. Does the Low-Rank Assumption limit the generality of method, making it effective only for a specific subset of (simpler) time series datasets? 2.If we encounter a complex trend in the future that Instance Normalization can't flatten, does your Root Purge method just fail? How can you convince us that it's still robust in that scenario? Lightly AI-edited
Characteristic Root Analysis and Regularization for Linear Time Series Forecasting Soundness: 3: good Presentation: 3: good Contribution: 3: good Rating: 6: marginally above the acceptance threshold Confidence: 4: You are confident in your assessment, but not absolutely certain. It is unlikely, but not impossible, that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. This paper advances a linear time-series forecasting framework centered on characteristic roots: in the noise-free regime, it shows that characteristic roots govern long-term dynamics; in noisy settings, it exposes the MSE training data-scaling law and the risk of spurious roots, indicating that simply enlarging the sample size is insufficient for robust generalization. Accordingly, the authors introduce two complementary structural regularization strategies: (i) low-rank constraints on the parameter matrix via Reduced-Rank Regression (RRR) and Direct Weight Rank Reduction (DWRR) to recover low-dimensional latent dynamics; and (ii) Root Purge, an adaptive loss that learns a noise null space and performs online root restructuring, with training applicable in either the time or frequency domain. Evaluated on benchmarks including Traffic, Electricity, Weather, Exchange, and ETT, the methods deliver strong results. 1. According to the reported results, the proposed methods achieve strong empirical performance across multiple datasets, which is particularly notable given the simplicity of the designs. 2. The paper provides substantial theoretical analysis with clear exposition to clarify the core concepts. 3. A clear roadmap is presented, and overall readability is high. 1. The core insight “preserving dominant characteristic roots to suppress noise” is not novel and aligns with fundamental conclusions from classical Singular Spectrum Analysis (SSA). 2. The technical contributions of DWRR and Root Purge appear simple. 3. The Root Purge objective may induce model degeneration: the model could prefer mapping any input to zero, and the paper offers no theoretical guarantees to prevent or mitigate this failure mode. 4. The theoretical analysis depends on i.i.d. Gaussian noise and stationarity assumptions to justify data efficiency and the regularizer; these assumptions often do not hold in practice, and no generalization beyond them is provided. 1. Under complex, non-stationary noise typical of real-world series, do the proposed methods retain theoretical advantages? 2. Can the approach be extended to more complex deep learning architectures? Lightly AI-edited
PreviousPage 1 of 1 (3 total rows)Next