|
CoLaP: Contrastive Learning with Adaptive Prompts for Continual Learning |
Continual learning (CL) aims to enable models to learn a sequence of new tasks without forgetting previously acquired knowledge. Prompt-based approaches, which adapt small prompt parameters while keep... |
2.00 |
13% |
See Reviews |
View AI Dashboard |
|
Teaching LLMs to Plan: Logical Chain-of-Thought Instruction Tuning for Symbolic Planning |
Large language models (LLMs) have demonstrated impressive capabilities across diverse tasks, yet their ability to perform structured symbolic planning remains limited, particularly in domains requirin... |
4.50 |
27% |
See Reviews |
View AI Dashboard |
|
HT-Transformer: Event Sequences Classification by Accumulating Prefix Information with History Tokens |
Deep learning has achieved strong results in modeling sequential data, including event sequences, temporal point processes, and irregular time series. Recently, transformers have largely replaced recu... |
2.40 |
11% |
See Reviews |
View AI Dashboard |
|
3DPhysVideo: 3D Scene Reconstruction and Physical Animation Leveraging a Video Generation Model via Consistency-Guided Flow SDE |
Video generative models have made remarkable progress, yet they often yield visual artifacts that violate grounding in real-world physical dynamics. Recent works such as PhysGen3D tackle single image-... |
5.00 |
3% |
See Reviews |
View AI Dashboard |
|
Latency-Aware Contextual Bandit: Application to Cryo-EM Data Collection |
We introduce a latency-aware contextual bandit framework that generalizes the standard contextual bandit problem, where the learner adaptively selects arms and switches decision sets under action dela... |
4.00 |
14% |
See Reviews |
View AI Dashboard |
|
TheMCPCompany: Creating General-purpose Agents with Task-specific Tools |
Since the introduction of the Model Context Protocol (MCP), the number of available tools for Large Language Models (LLMs) has increased significantly. These task-specific tool sets offer an alternati... |
5.00 |
0% |
See Reviews |
View AI Dashboard |
|
WAVE: Learning Unified & Versatile Audio-Visual Embeddings with Multimodal LLM |
While embeddings from multimodal large language models (LLMs) excel as general-purpose representations, their application to dynamic modalities like audio and video remains underexplored. We introduce... |
6.00 |
12% |
See Reviews |
View AI Dashboard |
|
CAP: Improving the Robustness of LLM-as-a-Judge Against Adversarial Score Manipulation via Comparative Augmented Prompting |
Large language models (LLMs) have been widely adopted in automated evaluation tasks, demonstrating human-aligned assessment capabilities. However, studies reveal that LLM-as-a-Judge systems exhibit si... |
3.33 |
7% |
See Reviews |
View AI Dashboard |
|
Structuring Hidden Features via Clustering of Unit-Level Activation Patterns |
We propose a self-supervised learning framework that organizes hidden feature representations across layers, thereby enhancing interpretability. The framework first discovers unit-level structures by ... |
3.33 |
21% |
See Reviews |
View AI Dashboard |
|
variCOT: Variational Inference for Implicit Chain-of-Thought in Language Models |
Chain-of-Thought (CoT) reasoning dramatically improves language model performance but incurs significant computational overhead through sequential token generation. While implicit CoT methods promise ... |
2.50 |
80% |
See Reviews |
View AI Dashboard |
|
Holistic Prompting: Joint Reasoning with Reusable States and Shortcut Discovery |
Large Language Models (LLMs) have demonstrated significant capabilities in complex reasoning tasks, often employing frameworks like Tree of Thoughts (ToT) and Chain-of-Thought (CoT).
However, such me... |
3.50 |
0% |
See Reviews |
View AI Dashboard |
|
Fair Graph Machine Learning under Adversarial Missingness Processes |
Graph Neural Networks (GNNs) have achieved state-of-the-art results in many relevant tasks where decisions might disproportionately impact specific communities. However, existing work on fair GNNs oft... |
5.33 |
0% |
See Reviews |
View AI Dashboard |
|
Hierarchical Speculative Decoding through Training-Free Slim-Verifier |
Speculative decoding (SD) addresses the high inference costs of large language models by having lightweight drafters generate candidates for large verifiers to validate in parallel. Current draft-ver... |
2.67 |
22% |
See Reviews |
View AI Dashboard |
|
Single-Step Bidirectional Unpaired Image Translation Using Implicit Bridge Consistency Distillation |
Unpaired image-to-image translation has seen significant progress since the introduction of CycleGAN. However, methods based on diffusion models or Schrödinger bridges have yet to be widely adopted in... |
5.00 |
0% |
See Reviews |
View AI Dashboard |
|
Fundamental bounds on efficiency-confidence trade-off for transductive conformal prediction |
Transductive conformal prediction addresses the simultaneous prediction for multiple data points. Given a desired confidence level, the objective is to construct a prediction set that includes the tru... |
5.00 |
5% |
See Reviews |
View AI Dashboard |
|
Nested Hash Layer: A Plug-and-play Module for Multiple-length Hash Code Learning |
Deep supervised hashing is essential for efficient storage and search in large-scale image retrieval. Traditional models generate hash codes of a single length, but this creates a trade-off between ef... |
3.50 |
0% |
See Reviews |
View AI Dashboard |
|
Dynamic Experts Search: Enhancing Reasoning in Mixture-of-Experts LLMs at Test Time |
Test-Time Scaling (TTS) enhances the reasoning ability of large language models (LLMs) by allocating additional computation during inference. However, existing approaches primarily rely on output-leve... |
4.00 |
4% |
See Reviews |
View AI Dashboard |
|
Style2Shape: Image Style Guided 3D Shape Material Generation |
This paper presents Style2Shape, a novel framework for generating physically-based rendering (PBR) materials for 3D models from a single reference image. Unlike existing methods limited by the diversi... |
3.50 |
84% |
See Reviews |
View AI Dashboard |
|
Automating the Refinement of Reinforcement Learning Specifications |
Logical specifications have been shown to help reinforcement learning algorithms in achieving complex tasks. However, when a task is under-specified, agents might fail to learn useful policies. In thi... |
5.50 |
25% |
See Reviews |
View AI Dashboard |
|
STDDN: A Physics-Guided Deep Learning Framework for Crowd Simulation |
Accurate crowd simulation is crucial for public safety management, emergency evacuation planning, and intelligent transportation systems. However, existing methods, which typically model crowds as a c... |
6.00 |
19% |
See Reviews |
View AI Dashboard |
|
ACON: Optimizing Context Compression for Long-horizon LLM Agents |
Large language models (LLMs) are increasingly deployed as agents in dynamic, real-world environments, where success requires both reasoning and effective tool use. A central challenge for agentic task... |
4.00 |
4% |
See Reviews |
View AI Dashboard |
|
Learning Part-Aware Dense 3D Feature Field For Generalizable Articulated Object Manipulation |
Articulated object manipulation is essential for various real-world robotic tasks, yet generalizing across diverse objects remains a major challenge. A key to generalization lies in understanding func... |
6.00 |
0% |
See Reviews |
View AI Dashboard |
|
Capturing Gaze Shifts for Guidance: Cross-Modal Fusion Enhancement for VLM Hallucination Mitigation |
Vision language models (VLMs) often generate hallucination, i.e., content that cannot be substantiated by either textual or visual inputs. Prior work primarily attributes this to over-reliance on ling... |
4.00 |
4% |
See Reviews |
View AI Dashboard |
|
Rethinking the Value of Multi-Agent Workflow: A Strong Single Agent Baseline |
Recent advances in LLM-based multi-agent systems (MAS) show that workflows composed of multiple LLM agents with distinct roles, tools, and communication patterns can outperform single-LLM baselines on... |
4.50 |
33% |
See Reviews |
View AI Dashboard |
|
TusoAI: Agentic Optimization for Scientific Methods |
Scientific discovery is often slowed by the manual development of computational
tools needed to analyze complex experimental data. Building such tools is costly
and time-consuming because scientists m... |
3.60 |
0% |
See Reviews |
View AI Dashboard |
|
One Bad Sample May Spoil the Whole Batch: A Novel Backdoor-Like Attack Towards Large Batch Processing |
As hardware accelerators like TPUs and large-memory GPUs continue to evolve rapidly, an increasing number of Artificial Intelligence (AI) applications are utilizing extremely large batch sizes to acce... |
4.67 |
0% |
See Reviews |
View AI Dashboard |
|
DERMARK: A Dynamic, Efficient and Robust Multi-bit Watermark for Large Language Models |
As large language models (LLMs) grow more powerful, concerns over copyright infringement of LLM-generated texts have intensified. LLM watermarking has been proposed to trace unauthorized redistributio... |
4.50 |
9% |
See Reviews |
View AI Dashboard |
|
KineDiff3D: Kinematic-Aware Diffusion for Category-Level Articulated Object Shape Reconstruction and Generation |
Articulated objects, such as laptops and drawers, exhibit significant challenges for 3D reconstruction and pose estimation due to their multi-part geometries and variable joint configurations, which i... |
3.00 |
51% |
See Reviews |
View AI Dashboard |
|
Guaranteeing Conservation of Integrals with Projection in Physics-Informed Neural Networks |
We propose a novel projection method that guarantees the conservation of integral quantities in Physics-Informed Neural Networks (PINNs). While the soft constraint PINNs use to enforce the structure o... |
2.50 |
0% |
See Reviews |
View AI Dashboard |
|
Personalization Under Value Conflict: Resolving Contradictory Preferences with Paired Fine-Tuning |
Large language models (LLMs) are increasingly expected to capture not only broadly shared human universal values but also the diverse and often contradictory preferences of individual users. Existing ... |
3.00 |
24% |
See Reviews |
View AI Dashboard |
|
SPG: Sandwiched Policy Gradient for Masked Diffusion Language Models |
Diffusion large language models (dLLMs) are emerging as an efficient alternative to autoregressive models due to their ability to decode multiple tokens in parallel. However, aligning dLLMs with human... |
5.00 |
4% |
See Reviews |
View AI Dashboard |
|
EmotionThinker: Prosody-Aware Reinforcement Learning for Explainable Speech Emotion Reasoning |
Emotional information in speech plays a unique role in multimodal perception. However, current Speech Large Language Models (SpeechLLMs), similar to conventional speech emotion recognition (SER) syste... |
6.50 |
13% |
See Reviews |
View AI Dashboard |
|
TemMed-Bench: Evaluating Temporal Medical Image Reasoning in Vision-Language Models |
Existing medical reasoning benchmarks for vision-language models primarily focus on analyzing a patient's condition based on an image from a *single* visit. However, this setting deviates significantl... |
4.00 |
0% |
See Reviews |
View AI Dashboard |
|
Rotation Control Unlearning: Quantifying and Controlling Continuous Unlearning for LLM with The Cognitive Rotation Space |
As Large Language Models (LLMs) become increasingly prevalent, their security vulnerabilities have already drawn attention. Machine unlearning is introduced to seek to mitigate these risks by removing... |
4.00 |
0% |
See Reviews |
View AI Dashboard |
|
Latent Planning Emerges with Scale |
LLMs can perform seemingly planning-intensive tasks, like writing coherent stories or functioning code, without explicitly verbalizing a plan; however, the extent to which they implicitly plan is unkn... |
5.33 |
0% |
See Reviews |
View AI Dashboard |
|
Knowledge distillation through geometry-aware representational alignment |
Knowledge distillation is a common paradigm for transferring capabilities from larger models to smaller ones. While traditional distillation methods leverage a probabilistic divergence over the output... |
4.00 |
0% |
See Reviews |
View AI Dashboard |
|
Uncertainty‑Routed Human–LLM Curation and Calibration for ANLI |
Adversarial NLI (ANLI) reveals distribution-shift failures that static benchmarks
miss, motivating evaluation and curation that are explicitly uncertainty-aware.
We present URC2—Uncertainty-Routed Cur... |
2.00 |
34% |
See Reviews |
View AI Dashboard |
|
Controlling a $\mu$RTS agent using Decision Transformers |
Decision Transformers (DT) are a Return-Conditioned Supervised Learning (RCSL) technique. A DT policy predicts actions by attending to a limited history of tokens that encodes states, actions and retu... |
2.00 |
0% |
See Reviews |
View AI Dashboard |
|
Can Text-to-Video Models Generate Realistic Human Motion? |
Recent advances in text-to-video (T2V) generation have yielded impressive progress in resolution, duration, and prompt fidelity, with models such as Pika, Gen-3, and Sora producing clips that appear c... |
4.50 |
53% |
See Reviews |
View AI Dashboard |
|
Parameter-Efficient Fine-Tuning of LLMs with Mixture of Space Experts |
Large language models (LLMs) have achieved remarkable progress, with Parameter-Efficient Fine-Tuning (PEFT) emerging as a key technique for downstream task adaptation. However, existing PEFT methods m... |
4.00 |
18% |
See Reviews |
View AI Dashboard |
|
QuRL: Rubrics As Judge For Open-Ended Question Answering |
Reinforcement Learning from Verifiable Rewards (RLVR) has significantly improved the performance of large language models (LLMs) on tasks with gold ground truth, such as code generation and mathematic... |
5.33 |
13% |
See Reviews |
View AI Dashboard |
|
Improved Sample Complexity Bounds For Diffusion Model Training Without Empirical Risk Minimizer Access |
Diffusion models have demonstrated remarkable performance in generating high-dimensional samples across domains such as vision, language, and the sciences. Although continuous-state diffusion models h... |
3.50 |
0% |
See Reviews |
View AI Dashboard |
|
SCMF: Lightweight Retrieval-Augmented Generation via Retrieval Vector Compression |
With the widespread adoption of Retrieval-Augmented Generation (RAG) in knowledge-intensive tasks, efficiency bottlenecks become increasingly evident: storing and retrieving large-scale high-dimension... |
2.00 |
60% |
See Reviews |
View AI Dashboard |
|
Minimum-Excess-Work Guidance |
We propose a regularization framework inspired by thermodynamic work for guiding pre-trained probability flow generative models (e.g., continuous normalizing flows or diffusion models) by minimizing e... |
4.50 |
7% |
See Reviews |
View AI Dashboard |
|
Decoupled Classifier-Free Guidance for Counterfactual Diffusion Models |
Counterfactual generation aims to simulate realistic hypothetical outcomes under causal interventions. Diffusion models have emerged as a powerful tool for this task, combining DDIM inversion with con... |
3.50 |
10% |
See Reviews |
View AI Dashboard |
|
ParaS2S: Benchmarking and Aligning Spoken Language Models for Paralinguistic-aware Speech-to-Speech Interaction |
Speech-to-Speech (S2S) models have shown promising dialogue capabilities, but their ability to handle paralinguistic cues—such as emotion, tone, and speaker attributes—and to respond appropriately in ... |
5.50 |
0% |
See Reviews |
View AI Dashboard |
|
ChemReason: A Chemical Code-Driven Reasoning LLM via Verifiable Reinforcement Learning |
In chemistry, most research on large language models has centered on knowledge question answering and retrieval. However, these approaches fall short on core tasks such as open molecular generation an... |
2.00 |
7% |
See Reviews |
View AI Dashboard |
|
United Minds or Isolated Agents? Exploring Coordination of LLMs under Cognitive Load Theory |
Large Language Models (LLMs) exhibit a notable performance ceiling on complex, multi-faceted tasks, as they often fail to integrate diverse information or adhere to multiple constraints.
We posit... |
4.50 |
50% |
See Reviews |
View AI Dashboard |
|
Not All Clients Are Equal: Collaborative Model Personalization on Heterogeneous Multi-Modal Clients |
As AI becomes more personal, e.g., Agentic AI, there is an increasing need for personalizing models for various use cases.
Personalized federated learning (PFL) enables each client to collaboratively ... |
7.00 |
0% |
See Reviews |
View AI Dashboard |
|
Fairness Aware Reward Optimization |
LLMs are typically aligned with human feedback via reward models but demographic skews and group-dependent disagreements in annotations can propagate systematic unfairness. We introduce Fairness-Aware... |
3.00 |
4% |
See Reviews |
View AI Dashboard |