Few Batches or Little Memory, But Not Both: Simultaneous Space and Adaptivity Constraints in Stochastic Bandits
Analysis of space and adaptivity constraints in stochastic multi-armed bandits with memory limitations.
Analysis of space and adaptivity constraints in stochastic multi-armed bandits with memory limitations.
Manifold-orthogonal extrapolation method for physics-informed neural networks handling out-of-distribution regimes.
Dynamic curriculum learning approach using gradient-based difficulty estimation for adaptive example ordering.
GIP framework for selecting training examples for LLM fine-tuning by maximizing mutual information with task-specific signals.
IGU-LoRA: adaptive rank allocation for parameter-efficient fine-tuning of LLMs using integrated gradients and uncertainty scoring.
Federated unlearning framework reducing computation and communication costs via gradient conflict mitigation.
LLM-guided approach for interpretable dynamic graph clustering with semantic explanations of cluster formation and evolution.
Memory-efficient continual learning using prototypical exemplar condensation to reduce stored samples per class.
Temporal Aggregated Convolution method for efficient spiking neural network acceleration on SIMD hardware.
EVE: variational neuron with explicit prior and posterior for probabilistic computation at unit level.
Machine learning model using EEG data to predict treatment response for schizophrenia using vagus nerve stimulation.
Benchmark for training AI systems to plan and act in physical world with understanding of causal mechanisms and constraints.
Theoretical analysis extending Domingos interpolation formula to stochastic gradient descent, characterizing neural network generalization as kernel machines.
Deep metric learning approach (Siamese, Triplet, Vision Transformer) for identifying scribes in Chinese manuscript calligraphy datasets.
UVLM: unified framework for loading and benchmarking multiple vision-language models with standardized interface in Colab environment.
Self-training framework with label correction for handling noisy data in neural network training using bilevel optimization techniques.
Uses recurrent neural networks with spatio-temporal attention for traffic monitoring from acoustic fiber-optic sensors. Application-focused, limited AI interest.
FedPBS: federated learning algorithm addressing statistical heterogeneity and non-IID data for distributed ML training with privacy preservation.
Interpretable data augmentation for imbalanced learning that generates realistic, feasible samples with transparent procedures and adjustability.
Method for training 4-bit quantized CNNs on standard CPUs with PyTorch achieving full-precision accuracy parity for cost-effective deep learning.
CONSERVAttack method for testing high energy physics ML applications against physically motivated systematic uncertainties and adversarial robustness.
Chunk-Guided Q-Learning algorithm for offline reinforcement learning balancing bootstrapping error and policy flexibility over long horizons.
Aumann-SHAP framework for interaction-aware counterfactual explanations decomposing model transitions using cooperative game theory.
Benchmarking open-source PPG foundation models for biological age prediction, comparing task-specific vs general-purpose models across populations.
Gated graph attention networks for predicting duration of large-scale power outages induced by natural disasters.
Layer-attentive residuals and contrastive feature learning for mental health classification with overlapping diagnostic categories.
Bootstrap-based neural network inference framework for robust T2 relaxation distribution estimation in low-SNR pancreatic MRI imaging.
Analysis of redundant features emerging in Transformer next-token predictors, identifying gradient components responsible for seemingly useless feature computation.
Hyperbolic control mechanism using parallel transport to steer text-to-image models away from unsafe content generation.
ST-ResGAT spatio-temporal graph neural network for road pavement deterioration forecasting and predictive maintenance decision-support.
Framework for accelerating LLM inference using contextual sparsity predictors for ReGLU-based feed-forward networks with minimal accuracy loss.
Multifidelity surrogate modeling combining high and low-resolution simulations for nuclear reactor CFD analysis parameter exploration.
Analysis of training-inference mismatch in neural networks with soft vs hard selection, using logic gate networks as test case.
TACTIC method for tabular anomaly detection using in-context learning with foundation models, advancing unsupervised learning for anomaly detection tasks.
Hybrid architecture combining classical ML for customer segmentation with RAG-enabled LLMs for personalized financial services marketing content generation.
Gradient modulation and projection techniques balance optimization across modalities in multimodal domain generalization tasks.
Pocket-K AI-ECG system using ECGFounder foundation model for non-invasive hyperkalemia detection with handheld deployment.
Efficient procedure for evaluating excess risk of empirical risk minimization using black-box access with minimal data and compute.
Self-Indexing KVCache predicts sparse attention from compressed keys to reduce KV cache bottleneck in long-context LLM inference.
Addresses domain skew in federated learning through feature decoupling and calibration across distributed clients with diverse data.
GoldenStart improves flow-matching RL policies via Q-guided priors and entropy control for faster inference and better exploration.
Mathematical foundation for sampling Boltzmann distributions via normalizing flows, proving existence of transport map approximations.
Unified functional analytic framework interpreting supervised and unsupervised learning as variational optimization over function spaces.
SIREN auto-decoder framework for high-fidelity compression of seismic velocity models using implicit neural representations.
Spectral clipping optimization technique for LLM training that addresses spectral norm instability and gradient noise issues in standard optimizers.
Theoretical analysis of online convex optimization with time-varying constraints, providing regret and constraint violation bounds.
ECG-Reasoning-Benchmark evaluates step-by-step clinical reasoning in multimodal LLMs for ECG interpretation across 6,400+ samples.
Methods for localizing and editing knowledge in large audio-language models, addressing factual errors across acoustic and language modules.
Refold refines protein inverse folding by combining template-based methods with structural matching and fusion for amino acid sequence design.
DeLL framework addresses catastrophic forgetting in autonomous driving via Dirichlet process mixture models and front-door causal adjustment.