Interpretable Diagnostics and Adaptive Data Assimilation for Neural ODEs via Discrete Empirical Interpolation
Framework using Discrete Empirical Interpolation Method for interpretability in neural ODEs and dynamical system analysis.
Framework using Discrete Empirical Interpolation Method for interpretability in neural ODEs and dynamical system analysis.
Review of neural network approaches for precipitation prediction, comparing traditional NWP with deep learning methods.
Research on jointly learning sequential and relational data for prediction tasks involving entities, integrating sequence and graph modeling.
ML approach to optimize power line de-energization decisions for wildfire risk mitigation by solving mixed-integer linear programs faster for operational power systems.
Research applying model-based reinforcement learning to improve variable selection heuristics in branch-and-bound solvers for mixed-integer linear programming and combinatorial optimization problems.
Invertible flow-based method for learning data-driven manifolds in irregularly-sampled time series classification.
Study on impact of label quality versus model complexity in time series anomaly detection with limited labels.
Adversarial robustness evaluation of CSI-based wireless sensing models for human activity recognition.
Deep RL study for dynamic algorithm configuration using DDQN and PPO on evolutionary algorithm parameter control.
Large-scale evaluation of hybrid quantum-classical autoencoders for unsupervised network intrusion detection.
Unified positional encoding framework using group actions for transformers, unifying rotational and additive approaches.
Causal framework for interpretable and controllable generative models with theoretical guarantees.
Technique addressing output logit divergence instability during LLM pretraining via embedding centering.
One-shot reinforcement learning approach for improving LLM reasoning across multiple domains with minimal data.
Fairness-aware federated learning method with calibrated server updates across demographic groups.
Theoretical framework jointly optimizing source weights and transfer quantities in multi-source transfer learning.
Multigrade deep learning framework for structured error refinement in neural network training.
Pre-trained transformer for brain motor decoding that generalizes across recording sites and subjects.
Context-aware runtime monitors for safe AI-based autonomous systems using ensemble ML controllers.
Theoretical framework for online learning with partial feedback and collection version spaces.
Rate-distortion framework for lossy compression of transformer intermediate representations to reduce inference compute and memory.
Analysis of safety properties in diffusion-based LLMs versus autoregressive LLMs, showing robustness against jailbreak attacks.
Study of capability acquisition in transformers tracking geometric changes and linear probes across model scales and algorithmic task difficulty levels.
Quantitative selection theorems proving that strong task performance under uncertainty necessitates world models and belief-like memory structures in agents.
Efficient evaluation algorithm for excess risk of large-scale empirical risk minimization using black-box access with single dataset.
Machine unlearning approach using key deletion in model architecture for privacy-compliant data removal without requiring full training data access.
Adaptive stock price prediction framework using autoencoder-gated dual node transformers with RL control for regime-dependent market behavior.
Theoretical framework for RLHF with multi-source imperfect preferences, deriving regret bounds when feedback comes from multiple annotators with systematic mismatches.
Trillion-parameter scientific multimodal foundation model with advanced agent capabilities spanning 100+ scientific domains and general reasoning tasks.
Method for learning low-dimensional policy manifolds in reinforcement learning through state-occupancy matching to improve sample efficiency.
Novel GNN architecture using cross-attentive cohesive subgraph embedding to address oversquashing and improve information flow in graph neural networks.
Theoretical analysis of safety verification limits for self-improving AI systems, formalizing compatibility between bounded risk and unbounded utility.
Uses LLM to dynamically generate curriculum over actions for RL agents, progressively introducing complex actions during training.
Studies how deep networks assign higher density to simpler out-of-distribution data than in-distribution test data.
Multimodal representation learning framework for e-commerce product understanding combining reasoning with product attributes.
Prompt-based continual learning method for next activity prediction that handles concept drift without catastrophic forgetting.
Language agents that learn adaptive policies at test-time through environment interactions, improving performance via iterative refinement.
Data-driven framework using skeleton-based biomechanical analysis and motion modeling for personalized dart training.
Pipeline using street-view image analysis and ML imputation to extract building elevation data for flood risk assessment in Texas.
Federated learning approach using HAPS networks with weighted client selection to handle non-IID data distributions.
Causal k-means clustering algorithm to identify heterogeneous treatment effects across unknown subgroups.
Parameter-efficient adaptation method for foundation models via black-box visual prompting without full parameter access.
Framework for approximating probability distributions using weighted particles via maximum mean discrepancy and gradient flows.
Empirical study measuring how prompt and response characteristics impact energy consumption and inference costs of LLM operations.
Develops polynomial-time algorithm for solving Stampacchia variational inequalities under the Minty condition.
Proposes multi-timescale variants of primal-dual hybrid gradient algorithm for distributed optimization problems.
Deep learning framework for image zero-watermarking using noise-adversarial training to learn distortion-invariant features.
Introduces AICO, a tool for testing feature significance in supervised learning models to improve interpretability and fairness.
Presents method to train generative models that learn causally disentangled latent representations using context modules.
Proposes boosted quantile regression neural networks with entropy analysis for predicting patterns in complex dynamical systems.