ICLR 2026 - Reviews

SubmissionsReviews

Reviews

Summary Statistics

EditLens Prediction Count Avg Rating Avg Confidence Avg Length (chars)
Fully AI-generated 2 (50%) 3.00 3.00 2080
Heavily AI-edited 0 (0%) N/A N/A N/A
Moderately AI-edited 0 (0%) N/A N/A N/A
Lightly AI-edited 0 (0%) N/A N/A N/A
Fully human-written 2 (50%) 2.00 4.50 2330
Total 4 (100%) 2.50 3.75 2205
Title Ratings Review Text EditLens Prediction
Sparse-Smooth Decomposition for Nonlinear Industrial Time Series Forecasting Soundness: 2: fair Presentation: 3: good Contribution: 2: fair Rating: 4: marginally below the acceptance threshold Confidence: 4: You are confident in your assessment, but not absolutely certain. It is unlikely, but not impossible, that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. This paper proposes a novel framework, NL-CS3 (Nonlinear Causal Sparse-Smooth Soft Sensor), for industrial time series forecasting. The method decomposes the problem into two interpretable stages: (1) Causal Sparse-Smooth Feature Extraction (CSSFE), which selects relevant sensors and enforces smooth temporal dynamics via sparsity and Laplacian regularization; and (2) Nonlinear Causal Mapping (NCM), which applies a static nonlinear function (e.g., neural network) to predict the target variable. The paper provides theoretical guarantees for convergence, selection consistency, and generalization, and demonstrates superior empirical performance on a refinery process dataset compared to 13 baselines, including Transformers and LSTMs. 1、The paper targets a realistic and important problem in industrial soft sensing—combining interpretability, robustness, and nonlinear modeling—which is often overlooked in deep learning approaches. 2、The authors provide convergence proofs, selection consistency, and a generalization bound, which are rare for nonlinear sparse-smooth models. 3、The decomposition into interpretable sparse-smooth feature extraction and nonlinear mapping is conceptually clean and physically interpretable, aligning with Wiener system theory. 4、The framework identifies physically meaningful sensors and dynamic modes that correspond to known industrial phenomena, enhancing trust and usability. 1、All experiments are conducted on a single industrial refinery dataset. It remains unclear whether the method generalizes to other industrial or non-industrial time series domains. 2、The alternating eigenvalue–proximal updates may be computationally heavy for large-scale sensor networks, yet the paper lacks runtime or scalability analysis. 3、While theorems are formally presented, many assumptions (e.g., eigenvalue positivity, irrepresentability, independence) are strong and may not hold in practice. 4、The baselines are strong but lack comparison with more recent time-series models. Please see the weaknesses. Fully AI-generated
Sparse-Smooth Decomposition for Nonlinear Industrial Time Series Forecasting Soundness: 2: fair Presentation: 2: fair Contribution: 2: fair Rating: 2: reject Confidence: 4: You are confident in your assessment, but not absolutely certain. It is unlikely, but not impossible, that you did not understand some parts of the submission or that you are unfamiliar with some pieces of related work. This paper focuses on time series modeling for industrial time series. The proposed method unifies sparse sensor selection, smooth temporal modeling, causal feature learning, and nonlinear prediction capability. Experiments are carried out based on a industrial refinery catalytic reforming unit dataset. 1. Applying time series modeling to practical tasks, i.e., industrial sensor data analysis and forecasting. 2. Throratical analysis are provided to support the proposed method. 1. The statement of the research question is not clear. What are the main difficulties in modeling industrial time series data? Why can't the existing time series analysis methods be directly applied? The author needs to clearly and explicitly explain this point. 2. Experiments are insufficient. Only one dataset is used. Also, some recent baselines are missing, specifically the deep time series models, e.g., iTransformer, TimesNet, Timemixer. 3. Fig 4 and 5's resolution and quality need improvement. please refer to the weakness Fully human-written
Sparse-Smooth Decomposition for Nonlinear Industrial Time Series Forecasting Soundness: 1: poor Presentation: 1: poor Contribution: 1: poor Rating: 2: reject Confidence: 5: You are absolutely certain about your assessment. You are very familiar with the related work and checked the math/other details carefully. This paper proposes nonlinear causal sparse-smooth network for industrial time-series forecasting. Operationally, it involves causal sparse-smooth feature extraction (CSSFE) and nonlinear causal mapping (NCM); theoretically, it performs formal analysis for consistency and convergence. Experiments are conducted on a single private dataset where the proposed method outperforms Transformer. - This paper provides a detailed diagram to demonstrate their algorithm. - This paper defines optimization problem formally and detail the solution process. - This paper introduces the private dataset used in detail. Writing and Motivation. - Authors noted causality in this paper. For example, the model name is nonlinear causal sparse-smooth network. Causality is a statistical term with strict demand of formal analysis and assumptions. From the paper I do not see even the fundamental assumptions of causal analysis (e.g., unconfoundedness assumption, positivity assmption). Moreover, identification property is not analyzed or even discussed but it is the core of causal analysis methods. - The solution process and its analysis is highly relevant to ADMM. Clarification and reference are needed. - The conclusion should include a critical discussion of the limitations of the proposed method and, based on these, propose future research directions aimed at addressing the identified shortcomings. - The authors identify three challenges—correlated sensors (variables), complex nonlinear dynamics, and interpretability—and claim these are unique to industrial time-series forecasting (TSF). However, these issues are prevalent across general TSF and broader machine learning contexts, and thus may not be unique to industrial applications. - Furthermore, the connection between the proposed method’s components and the stated challenges remains unclear. The authors should explicitly explain how each component is designed to address a specific challenge and substantiate these claims with empirical evidence (e.g., ablation or hyperparameter studies) and theoretical justification. Experiments. - Reproducibility. The dataset used for experiments seems not available. The code is not provided either. - Coverage. The experiments rely solely on a private dataset, which limits the generalizability and comparability of the results. Time-series forecasting research typically requires evaluation on diverse public datasets—such as ETT, Electricity, and PEMS04—many of which are industrial in nature. Moreover, numerous public industrial datasets are readily available, as confirmed through a simple search by the reviewer (not necessarily concrete). The absence of experiments on these datasets raises concerns regarding the breadth and representativeness of the experimental design. - Baselines. The selected baselines (e.g., ridge regression, LASSO, SVR) are outdated and insufficient for benchmarking modern time-series forecasting methods. The most recent baseline included, the Transformer, became outdated before 2020. Between 2020 and 2025, deep learning has driven significant advancements in time-series forecasting, exemplified by models such as TimeBridge (2025), CycleNet (2025), iTransformer (2024), P-sLSTM (2025), MixMamba (2024), and SparseTSF (2025). Furthermore, large language model (LLM)-based approaches, including TEST and Time-LLM, have emerged for zero-shot and few-shot forecasting. These models should have been incorporated to align the study with the contemporary research landscape. Authors should also investigate these literature from the beginning of their research program. Please kindly see the weakness window. Fully human-written
Sparse-Smooth Decomposition for Nonlinear Industrial Time Series Forecasting Soundness: 2: fair Presentation: 2: fair Contribution: 1: poor Rating: 2: reject Confidence: 2: You are willing to defend your assessment, but it is quite likely that you did not understand the central parts of the submission or that you are unfamiliar with some pieces of related work. Math/other details were not carefully checked. This paper proposes NL-CS3, a framework that decomposes industrial time series forecasting into sparse-smooth feature extraction followed by nonlinear mapping. The method combines LASSO sparsity, smoothness regularization, and Wiener model structure with theoretical guarantees. Experiments on one industrial refinery dataset show improvements over baselines including deep learning methods. 1. Addresses practical industrial challenges including high-dimensional sensor data and interpretability requirements. 2.Two-stage decomposition (linear feature extraction + nonlinear mapping) provides principled balance between interpretability and modeling capacity. The core technical contribution is limited, essentially combining existing techniques (LASSO, fused LASSO, Wiener models) without fundamental algorithmic breakthroughs or novel insights. 1.The experimental validation is severely insufficient, relying on a single dataset with 5000 samples from one refinery. The deep learning baselines perform suspiciously poorly, raising questions about fair comparison and proper tuning. Critical comparisons with specialized industrial soft sensor methods are missing, and no statistical significance testing is provided. 2.The theoretical assumptions are unrealistic for practical applications. The irrepresentability condition rarely holds for highly correlated industrial sensors, while the beta-min condition requires prior knowledge of true coefficients, contradicting the goal of automatic selection. 3.Scalability concerns arise from the complex alternating optimization requiring careful tuning of multiple hyperparameters (λ₁, λ₂, ℓ, s), making practical deployment challenging. 4.The claims about interpretability and causal learning lack rigorous validation with domain experts. Reproducibility is poor due to unavailable datasets and insufficient implementation details. 1. Were the deep learning models properly adapted for multivariate time series forecasting, given their unexpectedly poor performance? 2. What is the computational scaling behavior with increasing sensor count and temporal window size? Fully AI-generated
PreviousPage 1 of 1 (4 total rows)Next