A solver-in-the-loop framework for end-to-end differentiable coastal hydrodynamics
AegirJAX: fully differentiable hydrodynamic solver for coastal wave propagation and inverse problems using end-to-end differentiable computing.
AegirJAX: fully differentiable hydrodynamic solver for coastal wave propagation and inverse problems using end-to-end differentiable computing.
Privacy-preserving structural dataset (CSA-Graphs) for child sexual abuse imagery classification research without releasing original images.
Multi-agent deep reinforcement learning for energy optimization in cell-free massive MIMO networks. Distributed antenna reconfiguration and sleep mode selection.
Introduces Dynamic Context Evolution (DCE) to address cross-batch mode collapse in LLMs, where repeated prompting causes output repetitiveness. Proposes principled framework beyond ad hoc deduplication.
T-STAR: tree-structured reinforcement learning for multi-turn LLM agents with self-rectification and grafting for sparse reward optimization.
DINO-QPM: lightweight interpretability adapter converting DINOv2 visual foundation model features into human-interpretable representations.
Performance-energy trade-off analysis for real-time 3D Gaussian Splatting on edge GPUs with controlled computational budgets.
Deep learning framework for tea leaf disease classification using CNN models with explainability and robustness evaluation.
ATOM Report: comprehensive adoption analysis of ~1.5K open language models showing Chinese models surpassing US counterparts in 2025.
TraceSafe-Bench: first comprehensive benchmark evaluating LLM guardrails efficacy in multi-step tool-use agent trajectories.
Efficient learned data compression via dual-stream feature decoupling to balance probability modeling with system latency.
k-server-bench: automated mathematical discovery challenge for finding potential functions in the k-server conjecture.
Study on robustness of attribution heatmap explanations in deep networks trained to predict human image authenticity judgments.
Theory and practice of scalable Gaussian process regression using nearest neighbours for massive dataset applications.
Systematic study analyzing impact of retrieval pipeline components on RAG-based medical question answering systems with LLMs.
Integration of DeePMD-kit neural network potentials into GROMACS for multi-GPU accelerated molecular dynamics simulations.
Study comparing compatibility of face embeddings across different deep neural network models for biometric applications.
CADENCE: adaptive depth estimation system for autonomous vehicles that dynamically scales neural network complexity for embedded processors.
AlignPrune: noise-robust dynamic data pruning method using loss trajectory alignment to preserve clean samples under label noise.
Convergence rate analysis for asynchronous Q-learning with polynomial stepsize under high-dimensional central limit theorem conditions.
Personalized RewardBench: novel benchmark for evaluating reward models' ability to capture individual user preferences in LLM alignment.
Methodology for measuring generative AI power consumption across data centers to address proprietary data gaps and infrastructure planning.
arXiv paper on motion-controlled video generation with disentangled control and motion causality for physical scene dynamics.
Elastic Test-Time Training method addressing catastrophic forgetting in long-context 3D reconstruction with plastic inference-time updates.
Comprehensive arXiv survey of generative AI covering LLM architectures, deployment protocols, and applications as of early 2026.
Framework for smooth optimization of explicitly regularized sparse objectives via Hadamard overparametrization, enabling gradient-descent compatible solvers.
Theoretical analysis showing single labels on more samples outperforms multiple labels per sample for binary classifier comparison under noisy label budgets.
Web-based ML platform for phishing email detection achieving F1=0.99 on public dataset with interpretability and robustness for deployment.
Analytic federated learning (AFL) paradigm enabling closed-form solutions for federated learning with pre-trained models in single-round training.
SleepNet and DreamNet: deep learning models for visual classification via feature enrichment and reconstruction with pre-trained encoders.
Matrix Profile extension for anomaly detection in multidimensional time series from real-world applications like sensor monitoring.
DROP: distributional reinforcement learning framework with asymmetric learning rates modeling optimistic/pessimistic dopamine neuron behavior.
Pseudo-probability unlearning method for efficient privacy-preserving machine unlearning with reduced computational overhead and residual information.
Self-supervised physics-informed neural network for real-time human pose and dynamics estimation from sparse IMU sensor configurations.
Inference-time scaling method for discrete diffusion language models via trajectory refinement without retraining for reward optimization.
Negative Binomial VAE extension with discrete count-based latents for improved biological plausibility over continuous VAE representations.
Method for quantitatively estimating target task performance from unsupervised pretext tasks in semi/self-supervised learning before full training.
LNN-PINN: physics-informed neural network framework with liquid residual gating for improved predictive accuracy on complex PDE problems.
In-context learning approach for AutoML pipeline optimization beyond hyperparameter tuning, incorporating fine-tuning and ensembling techniques.
PhISM: physics-informed deep learning architecture for unsupervised hyperspectral imaging using continuous basis functions for interpretable latent representations.
LoFT method for long-tailed semi-supervised learning using foundation models with parameter-efficient fine-tuning to improve pseudo-label quality.
MDP modeling framework extending policy graphs for multi-stage stochastic programs with decision-dependent uncertainty and statistical learning.
Technique for detecting when reasoning LLMs overthink by analyzing entropy after chain-of-thought to enable early exiting.
Bayesian machine learning potentials for molecular simulations with uncertainty quantification using equivariant message passing.
Method for efficiently computing Lipschitz constant estimates for neural networks using local information to improve robustness certification.
Approximate replicability framework for machine learning algorithms that remain stable under input resampling.
Spectral clustering alternatives to Laplacian with group fairness constraints for equitable cluster representation.
GIFT unifies GRPO, DPO, and UNA in reinforcement learning framework for LLM alignment combining group-relative sampling with implicit preference learning.
LoRA-DA provides data-aware initialization for low-rank adaptation via theoretical framework and asymptotic analysis for parameter-efficient fine-tuning.
Nirvana specialized generalist model with task-aware memory mechanism for domain adaptation while preserving broad LLM capabilities.