The Phase Is the Gradient: Equilibrium Propagation for Frequency Learning in Kuramoto Networks
Theoretical analysis proving phase displacement in Kuramoto oscillator networks equals gradient of loss for frequency learning.
Theoretical analysis proving phase displacement in Kuramoto oscillator networks equals gradient of loss for frequency learning.
Graph neural network with diffusion-contrastive learning for wind nowcasting in regions lacking dense observation networks.
Combines SAINT attention mechanism with tree-based models like XGBoost for improved employee attrition prediction on tabular HR data.
AI agents for optimizing community water distribution systems by scheduling pumps and valves to meet demands while minimizing energy in dynamic real-world environments.
Combines physics-informed neural networks with quantum feature mapping for battery state-of-health estimation across chemistries.
Proposes SGED-TCD framework for lag-resolved causal discovery in multivariate time series with applications to environmental data.
Presents VeriSpecGen for automatic formal specification synthesis from natural language using LLMs with traceability for code verification.
Introduces LIRA method to defend LLMs against jailbreaks, backdoors, and unlearning by training models to align instruction representation.
Proposes CARE-ECG, causal agent-based reasoning framework for explainable ECG interpretation combining LLMs with physiological structure.
Demonstrates membership inference attacks on ECG foundation encoders, exposing participation privacy risks in self-supervised pretraining.
Proposes physics-aware spiking neural networks for energy-efficient wearable IMU-based human activity recognition on edge devices.
Organizes diffusion model fundamentals from Langevin perspective, offering simplified mathematical framework for beginners.
Derives exact finite-sample variance decomposition for subagging ensembles, providing mathematical characterization of resampling ratios.
Proposes CodeQuant for quantizing mixture-of-experts models by combining clustering and quantization to handle outlier-induced errors.
Introduces PepBenchmark, standardized benchmark with datasets and protocols for peptide drug discovery machine learning.
Presents IceCache for memory-efficient KV-cache management in long-sequence LLMs via CPU offloading and selective GPU retention.
Proposes WaveMoE, a mixture-of-experts foundation model for time series forecasting using wavelet-enhanced frequency-domain information.
Proposes Profiled Sparse Networks with heterogeneous connectivity patterns, benchmarked on vision and tabular classification tasks.
Introduces ReadMOF framework using chemical nomenclature and pretrained language models for metal-organic framework property prediction.
Studies how reward hacking during RLHF fine-tuning degrades LLM calibration and uncertainty quantification despite improving helpfulness.
Explores online continual self-supervised learning with focus on stability-plasticity trade-off in models learning from unlabeled streaming data.
MoEITS: green AI approach for reducing computational burden of Mixture-of-Experts LLMs through simplification.
Machine unlearning method for removing training data influence without direct access to forget sets.
Spectral analysis of LoRA weight updates showing low-frequency dominance enables efficient parameter-efficient fine-tuning.
Federated learning framework for IoT networks with energy efficiency optimization for small-scale datasets.
Self-distillation method for multi-turn LLM agents using skill-conditioning to improve sample efficiency in reinforcement learning.
On-policy distillation method for LLM alignment with adaptive weighting based on signal quality and credit assignment.
Communication-efficient optimization method extending Muon for federated learning of large language models.
Revisits value modeling in LLM reinforcement learning using generative critics for improved credit assignment.
Transformer architecture that dynamically determines its own depth and width during training by pruning redundant heads.
Reinforcement learning benchmark for Pokemon Red game with long horizons, sparse rewards, and complex control mechanics.
Improved online covariance estimation for averaged SGD with minimax-optimal convergence rates via trajectory regression.
Theoretical framework explaining how transformers learn in-context via mirror descent over mixture of transition distributions.
Proposes readiness indices based on Task2Vec embeddings to predict federated learning performance before training.
Establishes first information-theoretic lower bounds for score query complexity in diffusion model sampling.
Graph neural network domain adaptation method using information bottleneck and online distillation for robustness to distribution shifts.
Theoretical analysis of in-context learning in transformers beyond stationary settings, explaining how models adapt without parameter updates.
Subset selection framework using optimal transport for prototype selection with better handling of minority classes.
Novel framework for hypergraph neural networks using PDE-inspired diffusion equations to address oversmoothing and improve message passing.
Theoretical analysis of continuous-time online learning with two-layer neural networks in diffusion environments, establishing regret bounds.
ML approach using physics-informed representations to detect dynamical instability in safety-critical systems described by differential equations.
Dual formulation approach for robust reinforcement learning under distribution shift, addressing instability in adversarial RL methods.
Scheduling algorithm for LLM inference with provable stability when decode lengths are unknown, addressing memory overflow challenges in production systems.
Theoretical result showing K-way energy probes in predictive coding networks reduce to softmax, explaining apparent richness of per-hypothesis energy signals.
Theoretical analysis of KL divergence stability under Gaussian perturbations for non-Gaussian distributions, applicable to OOD detection with flow-based generative models.
Rollout tree-based credit assignment method for multi-step agentic RL, leveraging implicit state overlap between group rollouts to avoid uniform advantage assignment.
Analysis of credit assignment in reinforcement learning with verifiable rewards using polarity-entropy decomposition to diagnose token update patterns in LLM reasoning.
Benchmark evaluating mechanistic interpretability methods under conditions where model explanations are absent, controlling for elicitation confounding effects.
Optimization technique for continual learning reducing computational overhead of C-Flat while maintaining ability to balance new and old task performance.
Method for detecting LLM hallucinations using counterfactual graph intervention to identify causal mechanisms, moving beyond passive signal-based classification approaches.