Ax Yaolun Zhang, Ruohui Wang, Jiahao Wang, Yepeng Tang, Xuanyu Zheng, Haonan Duan, Hao Lu, Hanming Deng, Lewei Lu 3/30/2026

EVA: Efficient Reinforcement Learning for End-to-End Video Agent

EVA: reinforcement learning framework for video agents using MLLMs with adaptive reasoning to handle long video sequences and temporal dependencies efficiently.

Ax Saswata Bose, Suvadeep Maiti, Shivam Kumar Sharma, Mythirayee S, Tapabrata Chakraborti, Srijitesh Rajendran, Raju S. Bapi 3/30/2026

AI Generalisation Gap In Comorbid Sleep Disorder Staging

Deep learning model for automated sleep staging shows poor generalization to clinical populations with comorbid sleep disorders; proposes iSLEEPS to address limitations.

Ax Tom Marty, Eric Elmoznino, Leo Gagnon, Tejas Kasetty, Mizu Nishikawa-Toomey, Sarthak Mittal, Guillaume Lajoie, Dhanya Sridhar 3/30/2026

A Compression Perspective on Simplicity Bias

Theoretical analysis of simplicity bias in neural networks using minimum description length principle and compression framework.

Ax Siqiao Xue, Zhaoyang Zhu, Wei Zhang, Rongyao Cai, Rui Wang, Yixiang Mu, Fan Zhou, Jianguo Li, Peng Di, Hang Yu 3/30/2026

QuitoBench: A High-Quality Open Time Series Forecasting Benchmark

QuitoBench open benchmark for time series forecasting covering eight trend-seasonality-forecastability regimes with regime-balanced dataset design.

Ax Eric Yocam, Varghese Vaidyan, Yong Wang 3/30/2026

H-Node Attack and Defense in Large Language Models

H-Node ANC mechanistic framework identifies and defends hallucination representations in transformer LLMs at individual hidden-state dimensions.

Ax Hao Liang, Zhengyang Zhao, Meiyi Qiang, Mingrui Chen, Lu Ma, Rongyi Yu, Hengyi Feng, Shixuan Sun, Zimo Meng, Xiaochen Ma, Xuanlin Yang, Qifeng Cai, Ruichuan An, Bohan Zeng, Zhen Hao Wong, Chengyu Shen, Runming He, Zhaoyang Han, Yaowei Zheng, Fangcheng Fu, Conghui He, Bin Cui, Zhiyu Li, Weinan E, Wentao Zhang 3/30/2026

DataFlex: A Unified Framework for Data-Centric Dynamic Training of Large Language Models

Unified framework for data-centric dynamic training of LLMs with consistent interfaces for data selection and reweighting.