Ax Tim R. Davidson, Benoit Seguin, Enrico Bacis, Cesar Ilharco, Hamza Harkous 4/1/2026

Reasoning-Driven Synthetic Data Generation and Evaluation

Reasoning-driven approach for generating synthetic multi-modal training data without manual prompts, addressing scarcity of specialized AI training datasets.

Ax Xue Jiang, Tianyu Zhang, Ge Li, Mengyang Liu, Taozhi Chen, Zhenhua Xu, Binhua Li, Wenpin Jiao, Zhi Jin, Yongbin Li, Yihong Dong 4/1/2026

Think Anywhere in Code Generation

Proposes adaptive reasoning allocation during code generation for LLMs, addressing limitations of upfront thinking approaches in handling code complexity.

Ax Minhyuk Seo, Seongwon Cho, Minjae Lee, Diganta Misra, Hyeonbeom Choi, Seon Joo Kim, Jonghyun Choi 4/1/2026

GenOL: Generating Diverse Examples for Name-only Online Learning

GenOL framework for online learning with only concept names (name-only setup) enabling real-time adaptation to data distribution shifts in continual learning scenarios.

Ax Zehua Pei, Ying Zhang, Hui-Ling Zhen, Tao Yuan, Xianzhi Yu, Zhenhua Dong, Sinno Jialin Pan, Mingxuan Yuan, Bei Yu 4/1/2026

PreMoE: Proactive Inference for Efficient Mixture-of-Experts

Training-free framework for compiling sparse Mixture-of-Experts variants with predicted expert utility metric for deployment optimization.

Ax Ivan Y. Tyukin, Bogdan Grechuk, Evgeny M. Mirkes, Alexander N. Gorban 4/1/2026

When fractional quasi p-norms concentrate

Theoretical analysis of concentration properties for fractional quasi p-norms in high-dimensional spaces.

Ax Johannes Exenberger, Sascha Ranftl, Robert Peharz 4/1/2026

Deep Polynomial Chaos Expansion

Classical polynomial chaos expansion technique for surrogate modeling and uncertainty quantification in physical simulation.

Ax Da Chang, Yongxiang Liu, Ganzhao Yuan 4/1/2026

On the Convergence of Muon and Beyond

Theoretical convergence analysis of Muon optimizer for matrix-structured parameters in neural network training.

Ax Ahmed A. Elhag, Arun Raja, Alex Morehead, Samuel M. Blau, Hongtao Zhao, Christian Tyrchan, Eva Nittinger, Garrett M. Morris, Michael M. Bronstein 4/1/2026

Learning Inter-Atomic Potentials without Explicit Equivariance

Transformer-based inter-atomic potential model for molecular simulations without explicit equivariance constraints.

Ax Mingzhi Chen, Taiming Lu, Jiachen Zhu, Mingjie Sun, Zhuang Liu 4/1/2026

Stronger Normalization-Free Transformers

Research on normalization-free transformer architectures using Dynamic Tanh as alternative to standard normalization layers.