Detection of adversarial intent in Human-AI teams using LLMs
Detection framework using LLMs to identify adversarial attacks against human-AI teams, covering data poisoning, prompt injection, and prompt engineering threats.
Detection framework using LLMs to identify adversarial attacks against human-AI teams, covering data poisoning, prompt injection, and prompt engineering threats.
Method for discovering time-varying causal networks in neural time series without assuming known causal structure a priori.
Optimization framework using jointly learned surrogate models for multi-objective optimization of neural dynamical systems in biophysical simulations.
Theoretical analysis of synchronization gaps in diffusion transformers using coupled Ornstein-Uhlenbeck systems to explain mode interaction hierarchies in the reverse process.
Study of sensitivity in compressed transformers across architectures, identifying which components degrade catastrophically vs. compress well, with formal bounds on error propagation.
Novel unsupervised method for long-term outlier prediction in time series data using outlier score modeling.
Analysis of content-based routing in hybrid recurrent-attention architectures through controlled experiments on multiple benchmarks.
CLT-Forge library for mechanistic interpretability of LLMs using cross-layer transcoders and sparse feature attribution graphs.
Deep attention-based sequential ensemble learning framework for BLE-based indoor localization in care facilities.
Comparative analysis of machine learning approaches for vehicle fuel consumption prediction using Motor Trend dataset.
Benchmarking physics-guided and deep learning models for air quality index forecasting on standardized datasets.
Cognitive science study on human decision-making persistence in bandit tasks using confidence-freeze theory.
Multimodal misinformation detection system addressing visual content manipulation in social media.
Semi-supervised text classification using self-training with pseudo-labels to improve deep classifier performance on unlabeled data.
Learnable sparse memory banks with chapter-based routing for scaling knowledge storage in Transformers without prohibitive attention costs.
Training-free visual token pruning framework for efficient Vision-Language Model inference through text-conditioned subspace reconstruction.
Disentangled multi-modal representation learning via VAEs for molecular property prediction in drug discovery and materials science.
Weakly supervised learning method for classification from bag-level label proportions using dual-proportion constraints.
Theoretical analysis of zeroth-order optimization training dynamics through Neural Tangent Kernel perspective for gradient-free neural network training.
Lightweight continual learning method using pruned adaptation modules for foundation models compared against recent FM-based approaches.
Investigation of plasticity loss in deep reinforcement learning with proposed Optimization-Centric Plasticity hypothesis explaining adaptation dynamics.
Sharpness-aware fine-tuning approach for diffusion models to reduce reward hacking in reinforcement learning from human feedback.
Online data selection method for GRPO reinforcement learning that reuses high-signal prompts to improve LLM reasoning training efficiency.
Bayesian sequential design framework combining active learning, multi-armed bandits, and distributed computing for black-box optimization.
Theoretical analysis of batch size effects in stochastic conditional gradient optimization methods.
Pretrained video diffusion model repurposed as differentiable physics simulator for urban wind flow prediction.
Analysis of mechanistic interpretability in VAEs across image and tabular data modalities using causal circuit analysis.
Amortized variational inference method for logistic regression with missing covariate data using VAE-based approach.
Federated learning framework for fine-tuning Mixture-of-Experts LLMs on distributed data with privacy preservation.
Comparative study of LSTM, Transformer, and hybrid architectures for symbolic music generation tasks.
Data-driven weather forecasting using deep learning with reduced computational requirements compared to existing models.
Neural network surrogates for uncertainty quantification in physical systems through interval propagation methods.
Research on world models using reaction-diffusion dynamics as alternative to Transformers for predicting future environment states with better spatial inductive bias.
Analysis of Bregman geometry in transformer representations, showing how stream separation improves steering methods.
Formal framework for defining and analyzing agency in AI systems through continuous representation and mesa-optimization dynamics.
AutoKernel: open-source autonomous agent framework for GPU kernel optimization using iterative search on PyTorch models.
vLLM Semantic Router architecture for LLM inference optimization covering routing, caching, safety, and adaptive mechanisms.
TIDE: post-training system with learned routers for per-token early exit in LLM inference, no retraining required.
PLR: method using Plackett-Luce ranking to efficiently reorder in-context learning examples without exhaustive search.
Algorithms for constrained online convex optimization with memory constraints and predictions.
Fairness improvement method using exponentiated gradient approach for multi-class classification tasks to mitigate bias.
Study of introspective awareness mechanisms in LLMs, investigating whether steering detection reflects genuine circuitry or shallow heuristics.
DSPA: inference-time method using sparse autoencoders for LLM preference alignment without weight updates, enabling mechanistic steering.
Off-policy evaluation methods for ranking systems using offline logged data, addressing bias in inverse propensity score estimators.
Research on how learning systems can converge to incorrect solutions when feedback reliability is unobservable, addressing theoretical issues in optimization.
Continuous relaxation method for partition-constrained subset selection with submodular objectives, improving query complexity over existing local-search approaches.
Develops differential-geometric framework accounting for parameter redundancy in shallow neural networks via quotient geometry to measure intrinsic predictor properties.
Addresses on-device ML inference bottleneck by optimizing feature extraction from user behavior sequences for low-latency mobile app execution.
Open-source Bayesian optimization model for concrete strength prediction and mix design optimization, applying ML to materials science with public datasets.
Derives sharper generalization error bounds for Transformer architectures using offset Rademacher complexity across single and multi-head, multi-layer variants.