Can Graph Foundation Models Generalize Over Architecture?
Studies whether graph foundation models can generalize across different GNN architectures and graph characteristics, revealing limitations in current approaches.
Studies whether graph foundation models can generalize across different GNN architectures and graph characteristics, revealing limitations in current approaches.
Compares robustness quantification and uncertainty quantification methods for assessing classifier prediction reliability under distribution shift.
Critical analysis of tabular data generation via probabilistic circuits, questioning progress claims and evaluation protocols in current benchmarks.
Evaluates robustness of climate foundation models under out-of-distribution shifts from unprecedented climate states.
Theoretical generalization bounds for physics-informed neural networks solving incompressible Navier-Stokes equations.
MsFormer transformer-based framework for predictive maintenance in industrial IoT environments with complex sensor data dependencies.
Reinforcement learning approach for training autoregressive image models with policy-based tuning optimizing quality and diversity simultaneously.
AI system for automated spectroscopy interpretation in scientific discovery, reducing human bias in spectral analysis.
Polaris framework enabling self-improving agents for small language models through policy repair via experience abstraction and code modifications.
Bayesian learning framework for designing drone-assisted AED delivery networks in emergency medical services.
Adaptive prompt routing mechanism for selecting appropriate LLM or generative model based on input prompts, balancing fidelity and diversity.
Mathematical framework for solving high-dimensional stochastic optimal control problems with long horizons using Schrödinger eigenfunction methods.
Sparse packing format and CUDA kernels leveraging unstructured sparsity in LLM feedforward layers to reduce computational costs and model size.
Tight upper bounds on sample complexity for multi-group learning using one-inclusion graph prediction strategy and bipartite matching.
Foundational theoretical framework for learning under regime variation where learner, memory state, and evaluation conditions evolve over time.
Offline reinforcement learning method using guided expectation-maximization for action selection from multimodal action distributions in fixed datasets.
Model-based reinforcement learning using neural ODEs and SDEs to capture stochastic dynamics in fully and partially observed environments.
End-to-end reinforcement learning framework for heterogeneous DAG scheduling with gap-aware generation enabling rapid schedule adaptation across environments.
Diffusion model for unconditional molecular generation using permutation symmetry on quotient manifolds to enforce invariance in point-cloud generation.
Mechanistic interpretability framework identifying and attributing safety circuits in LLMs responsible for alignment, jailbreak, and backdoor behaviors.
Comparative study of seven ML models (XGBoost, LSTM, CNN-LSTM, etc.) for hourly air temperature and humidity forecasting in Chongqing.
Off-policy value-based reinforcement learning framework for LLMs enabling improved data utilization and sample efficiency for long-horizon tasks.
Energy-based model for graph generation using transport-aligned sampling to improve efficiency and quality in discrete domain generation.
Length-aware scheduling method accelerating reinforcement learning training for LLMs by optimizing rollout phase efficiency during chain-of-thought generation.
Continual learning framework using mixture-of-experts with similarity awareness for data-efficient adaptation to new tasks with limited samples.
Computationally efficient reinforcement learning algorithm for linear function approximation in MDPs satisfying linear Bellman completeness.
Federated learning approach combining differential privacy and Byzantine robustness to protect against both data leakage and adversarial server attacks.
Deep learning method for estimating aerodynamic variables (velocity, angle-of-attack) from piezoelectric sensor measurements on aircraft structures.
Systematic evaluation of prompting strategies (zero-shot, few-shot, chain-of-thought) for chart question answering across GPT-3.5, GPT-4, and GPT-4o models on ChartQA dataset.
TIPS framework improves RL training for search-augmented LLMs via turn-level reward shaping, addressing sparse rewards and credit assignment in reasoning tasks.
Multi-agent reinforcement learning agents develop efficient private communication protocol; performance drops with human-comprehensible language enforced.
CHANRG benchmark reveals limited generalization of RNA secondary-structure prediction models. 170K structured RNA families dataset.
Quantitative assessment of reference retrieval errors from 5 LLM platforms on 2,000 medical literature references. Evaluates Grok-2, ChatGPT, Gemini, Perplexity, DeepSeek.
Theoretical paper on thermodynamic principles and computational costs of maintaining symbolic interpretability in AI systems.
Theoretical analysis of low-rank knowledge distillation for LLMs with convergence and generalization guarantees. Covers compression techniques for efficient deployment.
Quantum-enhanced graph neural network for network intrusion detection exploiting relational dependencies in network traffic.
Quantum federated autoencoder for anomaly detection in IoT networks using distributed learning without centralizing raw data.
Multimodal fusion framework for predicting synthetic lethality in cancer drug development. Domain-specific bioinformatics research.
Quantum Wasserstein GAN for de novo drug design using generative AI. Focuses on drug discovery rather than ML applications or tools.
Quantum computing approach for probabilistic modeling over permutation-structured data using super-exponential symmetric group Fourier transform speedup.
Framework for computational arbitrage in AI model markets where arbitrageurs allocate inference budget across competing providers to undercut pricing.
First system enabling fully homomorphic encryption for end-to-end mmWave radar sensing with composable FHE kernels for signal processing and ML inference.
Token-level analysis of distributional shifts during RLVR fine-tuning of LLMs, examining mechanisms underlying reasoning improvements.
Data-driven approach using memory-augmented neural networks to model fluid wake effects for autonomous aerial and aquatic robots.
Functional component ablation framework analyzing specialization in hybrid language models combining attention with state space models or linear attention.
Verifiable synthetic benchmark for LLM-based insider threat detection using deterministic simulation engine to maintain ground truth and cross-artifact consistency.
Differential privacy framework for RLHF fine-tuning that decouples reward learning to preserve user privacy in LLM preference-based training.
Systematic benchmark comparing four multi-agent LLM orchestration architectures for financial document processing with cost-accuracy tradeoffs and scaling strategies.
Method leveraging intermediate layer representations in LLMs via Inter-Layer Structural Encoders to improve task-specific predictions beyond final-layer features.
Active learning approach using Rashomon ensemble for interpretable decision tree induction with direct hypothesis space characterization.