Genomic Next-Token Predictors are In-Context Learners
Demonstrates in-context learning emerges organically in genomic sequence models trained with next-token prediction on DNA sequences.
Demonstrates in-context learning emerges organically in genomic sequence models trained with next-token prediction on DNA sequences.
Develops methods for provably safe ML model updates preventing catastrophic forgetting and alignment drift in dynamic environments.
Proposes data filtering method for cross-domain offline RL addressing dynamics misalignment between source and target domains.
Analyzes KL regularization estimators in RL training of LLMs, comparing bias-variance tradeoffs of different approximation methods.
VL-RouterBench benchmark for evaluating vision-language model routing systems with quality-cost tradeoff assessment at scale.
Evaluates feature-dependent noise in preference-based reinforcement learning with realistic noise patterns correlated to observations.
Proposes GIFT method reconciling SFT and RL post-training for Large Reasoning Models via Gibbs initialization to prevent distributional collapse.
Solves constrained optimization problems via gradient-based methods using hierarchical score-matching spaces to overcome local optima.
Proposes neural characteristic function approach for graph domain adaptation addressing distributional shifts without manual feature design.
Studies sequential prediction with option to abstain in semi-adversarial settings mixing adversarial and stochastic instances.
Creates deep surrogate model for blast wave prediction that generalizes to out-of-distribution urban scenarios using machine learning.
Develops federated causal representation learning for decentralized counterfactual reasoning across coupled industrial systems while preserving data privacy.
Introduces SpeedTransformer, a transformer-based model for detecting transportation modes from smartphone GPS speed data.
Proposes Hyperparameter Trajectory Inference to adjust neural network hyperparameters post-deployment without full retraining using optimal transport.
Studies how pretrained Vision-Language-Action models resist catastrophic forgetting during continual learning in robot policy training.
Reduces transformer KV cache by using low-dimensional keys for attention selection while maintaining high-dimensional values, achieving O(log N) dimensional compression.
JAWS improves neural PDE solvers' long-term rollouts using spatially-adaptive Jacobian regularization to prevent spectral blow-up and unphysical divergence.
Adaptive channel pruning technique reduces communication overhead in split learning by selectively transmitting intermediate feature representations.
MR-Search proposes meta-reinforcement learning with self-reflection for agentic search, enabling agents to adapt strategies across episodes and improve in-context exploration.
Method for embodied agents to autonomously discover symmetry group structure for disentangled representation learning without requiring prior knowledge of group properties.
Theoretical investigation of deep residual networks' approximation capacity in continuous dynamical systems, quantifying minimal time-horizons for diffeomorphism approximation.
OMNIFLOW is a multimodal agent combining LLMs with physics-grounded reasoning for scientific tasks involving PDEs, addressing hallucinations through cross-domain generalization.
PhasorFlow: open-source Python library for computing on unit circle using complex phasors and unitary wave interference gates.
Time reparameterization technique for machine-learning reduced-order models of stiff dynamical systems improving training efficiency.
Reddit corpus annotated with moral sentiment and framing for NLP tasks related to moral language detection and analysis.
QFT: quantization-based approach for full-parameter fine-tuning of large language models with limited computational resources.
Reinforcement learning method for quantum circuit design handling device noise and connectivity constraints on real quantum hardware.
Byte-token enhanced language models for temporal point processes analysis to model event sequences with temporal dynamics and textual descriptions.
Method for improving mathematical reasoning in smaller LLMs by integrating arithmetic learning with knowledge distillation and data augmentation.
Survey of edge-cloud collaborative computing paradigms for distributed AI deployment, covering model optimization and LLM inference strategies.
Framework for constraint learning using pruned neural networks as tractable surrogates in optimization problems.
CIM-Explorer tool for optimizing binary and ternary neural networks on RRAM crossbar hardware architectures.
Information Imbalance metric for analyzing semantic information alignment in deep representations across text and image models.
Methods for constructing confidence intervals and hypothesis tests for functionals derived from online/sequential algorithms with computational constraints.
BiomedSQL benchmark for evaluating text-to-SQL systems on biomedical knowledge bases requiring implicit domain reasoning and scientific understanding.
Study on how prompt variability affects LLM code generation quality and functionality across different user backgrounds and expertise levels.
Deep learning models integrated with satellite data to reconstruct global forest carbon dynamics from 1988-2021 with uncertainty quantification.
Hebbian Physics Networks: self-organizing computational architecture using plastic transport geometry for solving physical dynamics problems.
SHAP-based framework for analyzing urban exercise inequality using spatial theory and machine learning on Shenzhen street data.
QR-learner model for estimating conditional treatment effects in trials using external data.
ToolRegistry: protocol-agnostic tool management library for function-calling LLMs, addressing fragmentation in tool integration.
Analysis of GNN generalization error to explain performance variance and benchmark skew in graph neural networks.
Benchmark study showing large multimodal models fail at inductive physical reasoning beyond training distribution.
EdiVal-Agent framework for automated, fine-grained evaluation of multi-turn image editing using object-centric assessment.
Detection methods for data contamination in RL post-training phase of LLMs, addressing evaluation validity gap.
Causal discovery method using multi-environment data to achieve full causal graph identifiability with minimal environments.
CBF-RL integrates Control Barrier Functions into RL training to enforce safety constraints during policy learning.
Unified optimization framework for jointly inferring time-varying network topologies and imputing missing graph signal data.
Neighbor GRPO extends Group Relative Policy Optimization to flow matching models with contrastive ODE-based approach for generative model alignment.
Knowledge Immunization Framework for selective knowledge erasure from LLMs via representation-aware activation signatures, addressing GDPR and safety.