Nonmyopic Global Optimisation via Approximate Dynamic Programming
Approximate dynamic programming approach for global optimization of expensive black-box functions as alternative to Gaussian process Bayesian optimization.
Approximate dynamic programming approach for global optimization of expensive black-box functions as alternative to Gaussian process Bayesian optimization.
Projection-free algorithms for online convex optimization with time-varying adversarial constraints and regret bounds.
Theoretical analysis of how iteration order affects convergence and stability in deep neural network training without learning rate schedules.
Methodological commentary on robust predictive modeling under distribution shifts in real-world deployment scenarios.
Task Tokens method adapts behavior foundation models to specific tasks via learnable tokens while preserving zero-shot generalization capabilities.
FastCache accelerates Diffusion Transformer inference through learnable linear approximation and spatial-aware token selection for hidden-state caching.
Defends RAG systems against knowledge poisoning attacks by detecting and mitigating adversarial text injections in external knowledge sources.
Masked training approach for robust arrhythmia detection from digitalized ECG images with temporal asynchrony and missing signal segments.
PepThink-R1 integrates LLMs with chain-of-thought supervised fine-tuning and reinforcement learning for interpretable cyclic peptide design optimization.
Generative model for molecular dynamics trajectories using Markov State Models to accelerate computational protein simulations.
LLMs perform automatic wireless modulation classification via discretized self-supervised candidate retrieval, avoiding distribution shift issues of supervised models.
Control-theoretic framework for LLM activation steering with feedback controllers, connecting empirical steering methods to proportional control theory for safety alignment.
NeST-BO proposes Newton-step targeting Bayesian optimization using Gaussian processes to learn gradient and Hessian information for expensive black-box problems.
Analyzes cryptanalytic model extraction attacks on ReLU-based DNNs with hard-label oracle access and polynomial-time complexity.
Sequence-level TopK (SeqTopK) improves Mixture-of-Experts routing in LLMs by adapting expert assignment per sequence rather than per token without retraining.
Cascading Bandits analyzes decision-making policies for edge inference with multiple models, providing theoretical regret guarantees for Explore-then-Commit and Thompson Sampling approaches.
LiteCache optimizes KVCache memory management for LLM inference using GPU-centric query similarity-driven approach to reduce memory overhead and improve CUDA Graph execution.
Repulsive Bayesian Prompt Learning addresses overfitting in prompt learning for foundation models using Bayesian inference framework for improved out-of-distribution generalization.
Balanced Fine-Tuning aligns LLMs with biomedical knowledge through confidence-weighted token-level optimization and adaptive reward mechanisms.
FedRE proposes a representation entanglement framework enabling federated learning across clients with heterogeneous model architectures and data.
SonicMoE optimizes Mixture of Experts inference through IO-aware and tile-aware techniques for high-granularity, sparse MoE language models.
Deep learning approach for radio path loss prediction in 5G networks with improved generalization across multi-transmitter scenarios and distribution shifts.
Concurrent training enhancements for Kolmogorov-Arnold networks using Newton-Kaczmarz method with FPGA implementation for improved efficiency.
Dual-State Action Pair (DSAP) primitive couples stochastic LLM generation with deterministic verification for reliable code generation agents.
Analyzes decentralized federated learning convergence with user mobility and data heterogeneity in next-gen wireless networks.
Provides theoretical framework explaining why diffusion models prefer direct data prediction over noise/velocity prediction in high-dimensional settings.
Extends Puzzle neural architecture search to reasoning LLMs, producing gpt-oss-puzzle-88B through MoE expert pruning and inference optimization.
Combines low-rank adaptation with quantization-aware unlearning to ensure LLM knowledge removal survives post-training 4-bit quantization.
Golden Layers method improves LLM knowledge editing via layer gradient analysis to identify optimal depth for updating model predictions per query.
cc-Shapley extends Shapley values for multivariate feature importance by incorporating causal context to address spurious associations.
TRC² architecture for continual learning in LLMs preventing catastrophic forgetting through decoder-only thalamic routing of cortical columns.
Web-Knowledge-Web pipeline iteratively crawls domain sources and knowledge graphs to discover small/medium enterprise suppliers with improved database coverage.
Establishes theoretical connection between drifting generative dynamics and Sinkhorn divergence-induced gradient flows with cross-minus-self decomposition.
AgentTrace framework for post-hoc root cause analysis in deployed multi-agent systems via causal graph reconstruction from execution logs.
Exploits massive redundancy in gradient transport to reduce real-time recurrent learning computational cost from O(n^4) via random sparsity patterns.
Connects adversarial robustness and LLM hallucinations through shared geometric principle formalized as Neural Uncertainty Principle with irreducible uncertainty bounds.
Benchmarks physics-guided and deep learning models for air quality index forecasting on region-specific datasets.
mSFT algorithm addresses overfitting in multi-task supervised fine-tuning by dynamically adjusting data mixture ratios based on task-specific learning dynamics.
Decouples exploration from policy optimization in RL using uncertainty-guided tree search for efficient autonomous exploration without intrinsic motivation.
Online learning algorithm balancing regret guarantees in adversarial/stochastic settings with safety constraints via COMPASS-Hedge method.
Architecture for aircraft health monitoring balancing accuracy and computational constraints under class imbalance and environmental uncertainty.
Deep learning approach for automated sleep staging in stroke patients with analysis of generalization gaps in clinical populations using Grad-CAM interpretations.
Method for steering code LLMs toward specific programming languages and libraries by manipulating activation space directions at inference time, tested on five language/library pairs across three open-weight models.
Analysis of response homogenization in RLHF-aligned LLMs showing reduced uncertainty estimation and implications for sampling.
Multimodal fusion approach for microservice incident detection handling missing modalities without static imputation.
Uncertainty-guided rebalancing technique for safety monitoring in cyber-physical systems with imbalanced time-series data.
Analysis of generalization in audio deepfake detection across datasets and model architectures.
Actor-critic reinforcement learning approach combining trajectory optimization with Sobolev learning for optimal control.
Knowledge-guided pretraining framework for multimodal foundation models applied to remote sensing applications.
Reproducibility analysis of 10 graph-based neural recommender papers from SIGIR 2022 assessing methodology and impact.