HyperMem: Hypergraph Memory for Long-Term Conversations
HyperMem: hypergraph-based memory architecture for conversational agents enabling long-term context tracking and high-order associations.
HyperMem: hypergraph-based memory architecture for conversational agents enabling long-term context tracking and high-order associations.
Quantum-inspired ARIMA methodology combining quantum autocorrelation with variational quantum circuits for time series analysis.
Vision-language benchmark (CrashSight) for evaluating traffic crash scene understanding from infrastructure perspective.
Physics-aligned simulator (SIM1) for generating synthetic data in deformable object robotic manipulation tasks.
Framework combining LLMs with graph neural networks for text-attributed graph learning in low-resource settings using GNN feedback.
Bayesian optimization method (MG-TuRBO) for high-dimensional traffic simulation calibration, comparing genetic algorithms with Bayesian approaches.
QuanBench+ unified benchmark for LLM quantum code generation across Qiskit, PennyLane, Cirq with 42 aligned executable tasks.
Benchmark evaluating robustness of LLM reasoning with 14 perturbation techniques applied to mathematical reasoning tasks.
Silhouette loss function for learning discriminative representations with explicit geometric properties in embedding space.
Distillation framework compressing genomic foundation models for efficient mRNA representation learning.
Quantum-classical hybrid molecular generator using VAE and quantum computing for interpretable drug discovery.
DRTO combines token-level RLHF with distributional robustness to improve LLM resilience to input perturbations and formatting changes.
Automated label function generation for data annotation using LLMs with structured exploration-exploitation strategy.
Analysis of cross-modal alignment between vision and language encoders using functional map framework from computational geometry.
TinyML Z-score anomaly detection system running on resource-constrained microcontrollers using power side-channel data.
Dual-branch reconstruction method for multivariate time series anomaly detection using autoregressive flow-based density estimation.
CSAttention: sparse attention mechanism for accelerating LLM inference by reducing KV-cache bottlenecks through centroid-scoring without retraining.
Flow-matching generative model for CFD surrogate modeling on unstructured meshes as alternative to deep learning approaches.
Framework evaluating when LLMs should act versus escalate decisions using uncertainty estimation across five real-world domains.
Machine learning system predicting user engagement in digital mental health interventions using explainable ML methods.
AlphaLab autonomous research system using frontier LLMs as agents to automate full experimental cycles in optimization domains without human intervention.
Analysis of hallucination phase transitions in Whisper ASR models using spectral sensitivity theorem and eigenspectra analysis.
Multi-task learning for wireless interference detection and identification using adversarial training methods.
Federated learning approach using exemplar replay to reduce catastrophic forgetting in continual learning with dynamic heterogeneity.
StructRL: recovers dynamic programming structure from distributional RL learning dynamics. Bridges data-driven and structured approaches for stable learning.
Bayesian inference for spiking neural networks in speech processing. Explores weight uncertainty and loss landscape smoothing for temporal tasks.
Evidential Transformation Network: adapts pretrained models for post-hoc uncertainty estimation. Efficient alternative to ensembles/MC dropout for deployed models.
VOLTA: benchmark comparing uncertainty quantification methods for deep learning. Evaluates 10 UQ baselines across modalities and distribution shifts.
Game-theoretic analysis of creator incentives in multi-agent recommender systems. Cooperative game formulation for fair collaboration in bandit problems.
PRAGMA: foundation models for banking event sequences. Transformer-based architecture with self-supervised pretraining on financial transaction data.
Skip-Connected Policy Optimization (SKPO) for reinforcement learning with reasoning tasks. Improves upon GRPO by addressing high-variance advantage estimation.
EvoLen: evolution-guided tokenization approach for DNA language models. Addresses fundamental tokenization design challenges in biological sequence modeling.
Experience replay for LLM post-training RL formalizing optimal buffer design as trade-off between sample efficiency and data freshness.
Tensor decomposition method quantifying uncertainty in LLM-based multi-agent systems accounting for communication and role dependencies.
CLOVER framework for multi-agent RL cooperation conditioning value decomposition on realistic wireless communication graphs.
LottaLoRA training paradigm showing frozen random backbones with trained LoRA adapters recover 96-100% performance across diverse tasks.
Adversarial sensor error framework for robust wind turbine fleet control against measurement errors and hacking.
Adaptive simulation experiment framework using pairwise comparisons to optimize LLM policies for operations management tasks.
Prompt optimization method decomposing reward variance into response and prompt variance to identify task amenability to optimization.
Multi-agent actor-critic reinforcement learning for disaster resilience controlling power, communication, and emergency response systems.
Evaluation metrics for SVG generation via element-level structural analysis using leave-one-out evaluation.
4-bit floating-point format (HiFloat4) for efficient language model pre-training on Ascend NPU hardware.
Guidance method for consistency models using joint flow distribution learning to enable classifier-free guidance without separate teacher model.
Training curriculum method for discrete flow-based image generation models to improve one-step sampling stability and quality.
Analysis of LoRA adapter spectral geometry to identify fine-tuning objectives and predict harmful model behavior in language models.
Safety steering mechanism for multimodal LLMs using dictionary-aligned concept control to prevent unsafe outputs without retraining.
Theoretical analysis of finite-sample properties and identifiability bounds for nonlinear Independent Component Analysis algorithms.
Survival analysis benchmark for predicting student dropout in learning analytics using OULAD dataset with dynamic and static representations.
Demonstrates power-law scaling of classification error with number of classes and how chain-of-thought decomposition reduces error through task splitting.
Temporal modeling framework for predicting student dropout using LMS data and logistic regression with counterfactual policy simulation.