S0 Tuning: Zero-Overhead Adaptation of Hybrid Recurrent-Attention Models
S0 tuning zero-overhead adaptation of hybrid recurrent-attention models outperforming LoRA on code generation.
S0 tuning zero-overhead adaptation of hybrid recurrent-attention models outperforming LoRA on code generation.
Function-based uncertainty quantification for safe learning-based control in safety-critical systems.
Learning to generate mixed quantum states prepared by shallow channel circuits in trivial phases.
RELISH lightweight architecture for text regression with LLMs using iterative latent state refinement.
Survey on Graph Neural Network acceleration techniques across algorithms, systems, and customized hardware.
RobustRAG defense framework with certifiable robustness against retrieval corruption attacks on RAG systems.
Inductive manifold learning approach for nonlinear dimensional reduction with local and global structure.
Domain adaptation with distribution shifts and unobserved confounding using linear structural causal models.
Topological Alignment Spectra method for analyzing multi-scale structural relationships in neural network representations.
Gaussian Process interpretation of wide neural networks with observation noise and arbitrary prior means.
Gradient-based hyperparameter learning via evidence lower bound objective from Bayesian variational methods.
Transformer-based decoder for Varshamov-Tenengolts codes correcting insertion, deletion, and substitution errors.
Multi-armed bandit algorithm using local graph structure to minimize regret under network interference.
Compositional automata learning technique for inferring models of concurrent systems through alphabet refinement.
SetONet neural operator for solving PDEs with variable sensor layouts by treating inputs as unordered sets.
Neural framework for learning conditional optimal transport maps using hypernetworks to generate adaptive transport parameters.
JUSSA framework uses steering vectors to improve LLM-as-a-judge reliability, detecting and mitigating sycophancy through honesty-promoting alternatives.
Binned semiparametric Bayesian networks for efficient kernel density estimation using data binning to reduce computational cost.
Double-Diffusion integrates ODE-prior with denoising diffusion models for spatio-temporal graph forecasting, balancing deterministic and stochastic components.
Klear-Reasoner model with long reasoning capabilities using gradient-preserving clipping policy optimization, with detailed training disclosures.
Knowledge component discovery in programming using representation learning on student code for personalized instruction systems.
Thompson sampling analysis for Sharpe ratio optimization in multi-armed bandit setting, addressing fractional objective with dependent mean-variance.
LSTM-based machine learning calibrator for agent-based epidemic models, learning inverse mapping from time series to SIR parameters.
EEG classification study comparing neural network architectures and optimizers across brain hemisphere frequency bands using TensorFlow/PyTorch.
Comprehensive survey of intrinsic dimension estimators under manifold hypothesis, reviewing theoretical foundations and comparing eight methods.
Analysis of weight constraints in linear smoothers for causal inference, balancing feature imbalance against parametric modeling assumptions.
Polychromic objectives approach to prevent mode collapse in reinforcement learning fine-tuning, preserving policy diversity during exploration.
Convergence analysis for decentralized SGD with high-probability guarantees, removing restrictive assumptions on gradient bounds and noise.
Mathematical analysis of incoherence in goal-conditioned autoregressive models, studying policy improvement through fine-tuning with online RL.
Empirical study examining the R-Learner framework limitations for network causal inference with graph-dependent heterogeneous treatment effects.
Theoretical analysis of diffusion models on discrete state spaces, establishing convergence guarantees for masked and random walk dynamics.
Tomographic Quantile Forests (TQF) for nonparametric uncertainty quantification in multivariate regression tasks.
Meta-probabilistic modeling framework for discovering latent structure across collections of related datasets using probabilistic graphical models.
Research on learnable Gray-Wyner networks for disentangling common and task-specific information in computer vision.
SAU method for machine unlearning in sparse LLMs via gradient masking and importance redistribution for privacy.
Research showing activation steering vectors in LLMs are fundamentally non-identifiable with large equivalence classes.
FIRE method for reinitialization in continual learning that balances stability and plasticity in neural networks.
Research on Natural Hypergradient Descent for bilevel optimization using Fisher information matrix as Hessian surrogate.
Evaluation of scaling laws for Chemical Language Models on downstream molecular property prediction tasks.
Introduces adaptive backbone scaling framework for class incremental learning to balance plasticity and stability while reducing memory overhead.
Framework for learning interpretable nonlocal operator kernels from data for climate process modeling.
Standardized benchmark dataset for epitope-specific antibody design with unified evaluation metrics for generative methods.
Preconditioned optimization method using row-momentum normalization for scalable matrix-based neural network training.
Projection-free algorithm for contextual bandits achieving logarithmic regret with improved efficiency over Online Newton Step.
Skill routing system for LLM agents that identifies relevant skills from large ecosystems before planning or execution.
Framework using LLMs to automatically design reward programs for cooperative multi-agent RL systems with sparse task feedback.
DreamerAD uses latent world models for efficient RL in autonomous driving, compressing diffusion sampling 80x with visual interpretability.
ERL framework enabling LLM agents to self-improve through experiential learning from past interactions and reflective adaptation.
Neuro-symbolic method for process anomaly detection combining neural networks with domain knowledge from process mining.
Federated learning approach for livestock growth prediction addressing privacy concerns and limited datasets in agricultural applications.