Learning-based Sketches for Frequency Estimation in Data Streams without Ground Truth
Learning-augmented sketches for frequency estimation in data streams without ground truth labels, improving over traditional memory-constrained methods.
Learning-augmented sketches for frequency estimation in data streams without ground truth labels, improving over traditional memory-constrained methods.
Framework quantifying impacts of model personalization on prediction accuracy and explanation quality in high-stakes domains like healthcare.
Reevaluation of policy gradient methods (PPO) for imperfect-information games, questioning necessity of complex DRL algorithms based on fictitious play and CFR.
Analysis of optimal denoising in score-based generative models, comparing full-denoising vs half-denoising under data regularity assumptions.
MASS: adaptive subspace selection method for model merging, combining multiple fine-tuned models without training overhead while matching separate endpoint accuracy.
RL finetuning for text-to-multiview diffusion models to improve few-step generation quality, balancing per-view fidelity and cross-view consistency.
Dense associative memories framework showing emergence of diffusion models from memory storage, bridging memorization and generalization in neural networks.
VERINA benchmark for evaluating LLM code generation with jointly generated specifications and proofs, addressing correctness verification challenges.
Coded robust aggregation method for distributed learning resilient to Byzantine attacks, improving gradient aggregation in federated settings.
Graph model merging technique for combining GNN models pre-trained on different domains with distribution discrepancy to create generalized models.
Discrete flow matching approach for single-step retrosynthesis prediction in organic chemistry, generating diverse and accurate reaction predictions.
Knowledge graph embedding methods for feature learning on large-scale graphs as external knowledge for downstream ML tasks, optimizing beyond link prediction.
Dynamic weighting approach combining supervised fine-tuning and reinforcement learning for LLM post-training, unifying on-policy and off-policy learning paradigms.
Tool-augmented LLM agents trained with synthetic code environments via RL to improve generalization on tool-use tasks, addressing brittleness with new tools and unseen workflows.
LANCE: low-rank activation compression method for efficient on-device continual learning, reducing memory costs during backpropagation in resource-constrained environments.
NanoFlux: adversarial dual-LLM framework for generating targeted training data to improve reasoning, achieving strong results with <200 examples through competitive Attacker-Defender dynamics.
Attribution-Guided Decoding uses interpretability to improve LLM instruction-following and factual accuracy.
PolyGraph Discrepancy metric provides absolute performance measure for graph generative models.
Tree search guidance method for controllable graph generation with diffusion models.
Theoretical bounds connecting Jensen-Shannon and Kullback-Leibler divergences for representation learning.
Cluster-PFN extends Prior-Data Fitted Networks to Bayesian clustering with uncertainty quantification.
AGRAG improves graph-based RAG for LLMs by addressing hallucination, reasoning, and answer quality issues.
FedSDWC applies causal learning to federated learning for handling out-of-distribution data shifts.
MAVA accelerates masked auto-regressive diffusion inference for practical reinforcement learning applications.
Derives tail distribution bounds for regret in optimism-based reinforcement learning algorithms.
Empirical comparison of flow matching variants with diffusion models for privacy-preserving tabular data synthesis.
PRISM complex-valued encoder explores phase relationships in semantic representations of neural sequence models.
Transfer learning approach using DTW for stress-strain behavior prediction in additive manufacturing.
Method to identify and measure social biases in text-to-image diffusion models via automated prompt search.
Novel architecture for polyphonic music generation using structural inductive bias and Smart Embedding.
Domain-adapted LLM fine-tuned for educational QA in space weather and heliophysics.
TRACE framework uses autoregressive density estimation for causal discovery in single event sequences.
MetaDOAR meta-controller applies multi-agent reinforcement learning to large-scale cyber-network security games.
TRC² architecture enables LLMs to continually learn and adapt without catastrophic forgetting through specialized decoder design.
Framework combining ML and contextual stochastic optimization for transit network design under demand uncertainty.
AOI framework enabling LLM agents to improve from failed cloud diagnosis trajectories in SRE automation with safety constraints.
KV cache optimization using low-dimensional attention selection to reduce transformer memory with O(log N) key dimensions.
Study examining many-shot prompting for test-time LLM adaptation, analyzing reliability and limits of in-context learning scaling.
Active feature acquisition method for biomedical applications optimizing measurement selection under temporal and cost constraints.
Theoretical analysis proving attention sinks are functionally necessary in softmax transformers for certain tasks.
Research addressing multimodal model underperformance in context-aided forecasting via improved context quality assessment.
Machine learning method using hypergraph pre-training to improve atrial fibrillation prediction in stroke patients.
Interactive benchmark environment for synthesizing flat-foldable origamis, testing AI systems' planning and causal reasoning in physical domains.
AI-ECG system using foundation models for non-invasive hyperkalemia detection and handheld deployment in clinical settings.
Neural compression framework using SIREN auto-decoders for high-fidelity compression of multi-structural seismic velocity models.
Graph-based verifier for LLM task planning that identifies and corrects hallucinations and flaws in agent-generated plans.
Multimodal framework combining weather foundation model with satellite data for fine-grained solar irradiance forecasting.
Post-hoc model-agnostic explanation method using informative perturbation selection for uncertainty-aware interpretability of black-box ML models.
Convergence analysis of Muon optimizer under heavy-tailed noise for nonconvex optimization in large-scale deep neural network training.
Analysis showing wider beam search in LLMs can degrade output quality due to overestimation bias in noisy scorer outputs, with theoretical grounding.