Ax Dung V. Nguyen, Hieu M. Vu, Nhi Y. Pham, Lei Zhang, Tan M. Nguyen 3/30/2026

Activation Steering with a Feedback Controller

Control-theoretic framework for LLM activation steering with feedback controllers, connecting empirical steering methods to proportional control theory for safety alignment.

Ax Tiansheng Wen, Yifei Wang, Aosong Feng, Long Ma, Xinyang Liu, Yifan Wang, Lixuan Guo, Bo Chen, Stefanie Jegelka, Chenyu You 3/30/2026

Route Experts by Sequence, not by Token

Sequence-level TopK (SeqTopK) improves Mixture-of-Experts routing in LLMs by adapting expert assignment per sequence rather than per token without retraining.

Ax R Sri Prakash, Nikhil Karamchandani, Sharayu Moharir 3/30/2026

Cascading Bandits With Feedback

Cascading Bandits analyzes decision-making policies for edge inference with multiple models, providing theoretical regret guarantees for Explore-then-Commit and Thompson Sampling approaches.

Ax Yassir Bendou, Omar Ezzahir, Eduardo Fernandes Montesuma, Gabriel Mahuas, Victoria Shevchenko, Mike Gartrell 3/30/2026

ReBaPL: Repulsive Bayesian Prompt Learning

Repulsive Bayesian Prompt Learning addresses overfitting in prompt learning for foundation models using Bayesian inference framework for improved out-of-distribution generalization.

Ax Zhenchao Tang, Fang Wang, Haohuai He, Jiale Zhou, Tianxu Lv, Jun Zhu, Shouzhi Chen, Minghao Yang, Yu Wang, Jiayang Wu, Yidong Song, Yaokun Li, Jiehui Huang, Dawei Huang, Zhi Song, Jianhua Yao 3/30/2026

Aligning LLMs with Biomedical Knowledge using Balanced Fine-Tuning

Balanced Fine-Tuning aligns LLMs with biomedical knowledge through confidence-weighted token-level optimization and adaptive reward mechanisms.

Ax Akhiad Bercovich, Nir Ailon, Vladimir Anisimov, Tomer Asida, Nave Assaf, Mohammad Dabbah, Ido Galil, Amnon Geifman, Yonatan Geifman, Izhak Golan, Roi Koren, Itay Levy, Zach Moshe, Pavlo Molchanov, Najeeb Nabwani, Mostofa Patwary, Omri Puny, Tomer Ronen, Itamar Schen, Elad Segal, Ido Shahaf, Oren Tropp, Ran Zilberstein, Ran El-Yaniv 3/30/2026

Extending Puzzle for Mixture-of-Experts Reasoning Models with Application to GPT-OSS Acceleration

Extends Puzzle neural architecture search to reasoning LLMs, producing gpt-oss-puzzle-88B through MoE expert pruning and inference optimization.

Ax Jo\~ao Vitor Boer Abitante, Joana Meneguzzo Pasquali, Luan Fonseca Garcia, Ewerton de Oliveira, Thomas da Silva Paula, Rodrigo C. Barros, Lucas S. Kupssinsk\"u 3/30/2026

Quantization-Robust LLM Unlearning via Low-Rank Adaptation

Combines low-rank adaptation with quantization-aware unlearning to ensure LLM knowledge removal survives post-training 4-bit quantization.

Ax Ping He, Om Khangaonkar, Hamed Pirsiavash, Yikun Bai, Soheil Kolouri 3/30/2026

Sinkhorn-Drifting Generative Models

Establishes theoretical connection between drifting generative dynamics and Sinkhorn divergence-induced gradient flows with cross-minus-self decomposition.

Ax Saswata Bose, Suvadeep Maiti, Shivam Kumar Sharma, Mythirayee S, Tapabrata Chakraborti, Srijitesh Rajendran, Raju S. Bapi 3/30/2026

AI Generalisation Gap In Comorbid Sleep Disorder Staging

Deep learning approach for automated sleep staging in stroke patients with analysis of generalization gaps in clinical populations using Grad-CAM interpretations.

Ax Nicolas M. M\"uller, Pavel Czempin, Franziska Dieckmann, Adam Froghyar, Konstantin B\"ottinger 3/30/2026

Does Audio Deepfake Detection Generalize?

Analysis of generalization in audio deepfake detection across datasets and model architectures.