Ax Duo Su, Huyu Wu, Huanran Chen, Yiming Shi, Yuzhu Wang, Xi Ye, Jun Zhu 4/6/2026

Diffusion Models as Dataset Distillation Priors

Dataset distillation method leveraging diffusion models as priors to synthesize compact, representative datasets with improved diversity and generalization.

Ax Junxiong Wang, Fengxiang Bie, Jisen Li, Zhongzhu Zhou, Zelei Shao, Yubo Wang, Yinghui Liu, Qingyang Wu, Avner May, Sri Yanamandra, Yineng Zhang, Ce Zhang, Tri Dao, Percy Liang, Ben Athiwaratkun, Shuaiwen Leon Song, Chenfeng Xu, Xiaoxia Wu 4/6/2026

When RL Meets Adaptive Speculative Training: A Unified Training-Serving System

Unified training-serving system combining RL with adaptive speculative decoding for accelerated LLM inference.

Ax Vignesh Gopakumar, Ander Gray, Dan Giles, Lorenzo Zanisi, Matt J. Kusner, Timo Betcke, Stanislas Pamela, Marc Peter Deisenroth 4/6/2026

Learning Physical Operators using Neural Operators

Physics-informed neural operators for solving PDEs with improved generalization beyond training distributions.

Ax Xiangyang Zhu, Yuan Tian, Qi Jia, Kaiwei Zhang, Zicheng Zhang, Chunyi Li, Kaiyuan Ji, Dongrui Liu, Zijian Chen, Lu Sun, Renrui Zhang, Yan Teng, Jing Shao, Wei Sun, Xia Hu, Yu Qiao, Guangtao Zhai 4/6/2026

SafeSci: Safety Evaluation of Large Language Models in Science Domains and Beyond

SafeSci: Framework for evaluating safety of large language models in scientific domains with comprehensive benchmarks.

Ax Aur Shalev Merin 4/6/2026

Temporal Credit Is Free

Recurrent network training without Jacobian propagation using hidden state temporal credit. Studies gradient normalization and online adaptation.

Ax Samuel Girard, Aurelien Bibaut, Arthur Gretton, Nathan Kallus, Houssam Zenati 4/6/2026

Fast Best-in-Class Regret for Contextual Bandits

Fast regret bounds for contextual bandits without realizability assumptions using pessimistic policy updates.