Score Distillation Beyond Acceleration: Generative Modeling from Corrupted Data
arXiv paper: Restoration Score Distillation framework for learning generative models from corrupted data. Novel ML research.
arXiv paper: Restoration Score Distillation framework for learning generative models from corrupted data. Novel ML research.
Foundation model for time series forecasting using mixture-of-experts architecture with decoupled training to handle diverse temporal patterns and multi-variable correlations.
Tutorial on diffusion and flow-based generative models covering mathematical foundations, ODEs, SDEs, and core algorithms for image, video, and multi-modal generation.
Offline reinforcement learning method for mismatched dynamics leveraging model-based approaches to explore high-reward states.
Self-organizing map extension addressing catastrophic forgetting in continual learning scenarios.
Privacy-preserving neural network framework integrating epidemiological modeling with differential privacy guarantees.
Learning to Reject framework extending ML models to abstain from predictions and explanations with low quality.
Method predicting better pre-trained weights by leveraging structural properties and retrodiction of forgetting.
Fast weight programmers with 2D matrix hidden states connecting RNNs, language modeling, and neurobiology.
Inverse game theory algorithm learning constraints from Nash equilibrium demonstrations using MILP formulations.
Transfer learning framework for EEG-based emotion recognition using domain-class prototypes.
Theoretical analysis of implicit regularization in diagonal linear networks via Lasso regularization path.
Tree-based group relative policy optimization for LLM agents addressing sparse supervision in multi-turn tasks.
Method aligning supervised fine-tuning with in-context learning activations to improve LLM generalization and calibration.
Offline reinforcement learning using in-context learning with linear Transformers for compositional Q-function estimation.
Parameter-efficient unlearning method for foundation models addressing privacy/safety with bounded weight growth.
Slow-Fast Policy Optimization framework for improving LLM reasoning via reinforcement learning with stable gradient updates.
Energy-based models combining classification and generation via adversarial training to improve SGLD stability.
Proposes continual low-rank adapters for LLM-based recommender systems handling evolving users and preferences without catastrophic forgetting.
Develops calibration-free soil sensing system using contrastive learning to predict moisture and macronutrients without retraining.
Presents augmentation-free graph contrastive learning via fractional-order neural diffusion networks for multi-scale structure learning.
Proposes standardized methodology for evaluating long-term sustainability and efficiency of ML models addressing Green AI gaps.
Demonstrates in-context learning emerges organically in genomic sequence models trained with next-token prediction on DNA sequences.
Develops methods for provably safe ML model updates preventing catastrophic forgetting and alignment drift in dynamic environments.
Proposes data filtering method for cross-domain offline RL addressing dynamics misalignment between source and target domains.
Analyzes KL regularization estimators in RL training of LLMs, comparing bias-variance tradeoffs of different approximation methods.
VL-RouterBench benchmark for evaluating vision-language model routing systems with quality-cost tradeoff assessment at scale.
Evaluates feature-dependent noise in preference-based reinforcement learning with realistic noise patterns correlated to observations.
Proposes GIFT method reconciling SFT and RL post-training for Large Reasoning Models via Gibbs initialization to prevent distributional collapse.
Solves constrained optimization problems via gradient-based methods using hierarchical score-matching spaces to overcome local optima.
Proposes neural characteristic function approach for graph domain adaptation addressing distributional shifts without manual feature design.
Studies sequential prediction with option to abstain in semi-adversarial settings mixing adversarial and stochastic instances.
Creates deep surrogate model for blast wave prediction that generalizes to out-of-distribution urban scenarios using machine learning.
Develops federated causal representation learning for decentralized counterfactual reasoning across coupled industrial systems while preserving data privacy.
Introduces SpeedTransformer, a transformer-based model for detecting transportation modes from smartphone GPS speed data.
Proposes Hyperparameter Trajectory Inference to adjust neural network hyperparameters post-deployment without full retraining using optimal transport.
Studies how pretrained Vision-Language-Action models resist catastrophic forgetting during continual learning in robot policy training.
Reduces transformer KV cache by using low-dimensional keys for attention selection while maintaining high-dimensional values, achieving O(log N) dimensional compression.
JAWS improves neural PDE solvers' long-term rollouts using spatially-adaptive Jacobian regularization to prevent spectral blow-up and unphysical divergence.
Adaptive channel pruning technique reduces communication overhead in split learning by selectively transmitting intermediate feature representations.
MR-Search proposes meta-reinforcement learning with self-reflection for agentic search, enabling agents to adapt strategies across episodes and improve in-context exploration.
Method for embodied agents to autonomously discover symmetry group structure for disentangled representation learning without requiring prior knowledge of group properties.
Theoretical investigation of deep residual networks' approximation capacity in continuous dynamical systems, quantifying minimal time-horizons for diffeomorphism approximation.
OMNIFLOW is a multimodal agent combining LLMs with physics-grounded reasoning for scientific tasks involving PDEs, addressing hallucinations through cross-domain generalization.
PhasorFlow: open-source Python library for computing on unit circle using complex phasors and unitary wave interference gates.
Time reparameterization technique for machine-learning reduced-order models of stiff dynamical systems improving training efficiency.
Reddit corpus annotated with moral sentiment and framing for NLP tasks related to moral language detection and analysis.
QFT: quantization-based approach for full-parameter fine-tuning of large language models with limited computational resources.
Reinforcement learning method for quantum circuit design handling device noise and connectivity constraints on real quantum hardware.
Byte-token enhanced language models for temporal point processes analysis to model event sequences with temporal dynamics and textual descriptions.