Ax Amos Goldman (NVIDIA Corporation), Nimrod Boker (NVIDIA Corporation), Maayan Sheraizin (NVIDIA Corporation), Nimrod Admoni (NVIDIA Corporation), Artem Polyakov (NVIDIA Corporation), Subhadeep Bhattacharya (NVIDIA Corporation), Fan Yu (NVIDIA Corporation), Kai Sun (NVIDIA Corporation), Georgios Theodorakis (NVIDIA Corporation), Hsin-Chun Yin (NVIDIA Corporation), Peter-Jan Gootzen (NVIDIA Corporation), Aamir Shafi (NVIDIA Corporation), Assaf Ravid (NVIDIA Corporation), Salvatore Di Girolamo (NVIDIA Corporation), Manjunath Gorentla Venkata (NVIDIA Corporation), Gil Bloch (NVIDIA Corporation) 3/25/2026

NCCL EP: Towards a Unified Expert Parallel Communication API for NCCL

NCCL EP, unified communication API for mixture-of-experts architectures in large language models built on NCCL with GPU-initiated RDMA.

Ax KT Tech innovation Group 3/25/2026

Mi:dm K 2.5 Pro

Mi:dm K 2.5 Pro, 32B parameter enterprise LLM supporting multi-step reasoning, long-context understanding, and agentic workflows in Korean and domain-specific applications.

Ax Weixuan Zeng, Pengcheng Wei, Huaiqing Wang, Boheng Zhang, Jia Sun, Dewen Fan, Lin HE, Long Chen, Qianqian Gan, Fan Yang, Tingting Gao 3/25/2026

OmniDiT: Extending Diffusion Transformer to Omni-VTON Framework

Diffusion Transformer framework for virtual try-on and try-off tasks combining both into unified model for fashion applications.

Ax Yan Xie, Tiansheng Wen, Tangda Huang, Bo Chen, Chenyu You, Stefanie Jegelka, Yifei Wang 3/25/2026

Scaling Attention via Feature Sparsity

Sparse Feature Attention method reducing transformer self-attention complexity through feature-level sparsity instead of sequence-level sparsity.