Experiential Reflective Learning for Self-Improving LLM Agents
Experiential Reflective Learning framework enabling LLM agents to self-improve by leveraging past interactions and adapting to specialized environments.
Experiential Reflective Learning framework enabling LLM agents to self-improve by leveraging past interactions and adapting to specialized environments.
Mechanistic interpretability analysis of how LLMs verbalize confidence scores versus actual accuracy using linear probes and activation steering.
Neuro-symbolic approach combining neural networks with domain knowledge for process anomaly detection in event logs.
Vision2Web: Hierarchical benchmark for evaluating AI agents on website development tasks from UI-to-code to full-stack implementation.
CarbonEdge: Carbon-aware deep learning inference framework for edge computing optimizing environmental impact alongside latency and throughput.
CDH-Bench: Benchmark evaluating vision-language models' commonsense-driven hallucinations when visual evidence conflicts with common sense.
LG-HCC proposes geometry-aware compression for 3D Gaussian Splatting to reduce storage overhead while maintaining rendering quality.
HISA improves efficiency of sparse attention mechanisms by optimizing hierarchical indexing to reduce bottlenecks in token-level key selection for LLMs.
MemFactory: unified inference and training framework for agent memory integration with RL optimization of memory operations in LLM agents.
FigAgent: multi-agent framework for automatic method illustration generation in AI papers via drawing middleware orchestration.
Reduced density matrix method from quantum chemistry for predicting and interpreting phase transitions during deep learning model training.
Optimizer-aware gradient-based online data selection framework for sequential LLM fine-tuning with step-dependent sample utility estimation.
Personalized federated fine-tuning approach for language models on distributed heterogeneous task datasets with improved generalization.
Evolution strategies for Deep RL pretraining offering derivative-free, computationally efficient alternative to standard deep reinforcement learning.
Continual learning framework for resource-constrained agents using stochastic bridge diffusion process for temporal memory management.
Perspective on sustainability challenges in AI-driven molecular and materials discovery across QM data, training, and automation pipelines.
Empirical evidence that classifier-based safety gates fail for self-improving AI systems across multiple model architectures.
Symbolic mixture-of-experts model for predicting cross-location hurricane evacuation behavior with population-level adaptation.
LLM-based curriculum learning framework for reinforcement learning agents applied to Blackjack game strategy.
Sparse interpretable machine learning models for improving branching decisions in mixed-integer programming solvers without GPU requirements.
Open-source adaptive router for multi-model LLM serving using cost-aware contextual bandits with non-stationary pricing and quality changes.
Epileptic seizure detection from EEG signals using graph convolutional neural networks on frequency band features.
Sit-to-stand transition detection using smart lacelock sensor for fall risk assessment in older adults.
Normalizing flow models using Lévy process distributions for heavy-tailed financial risk modeling.
Offline reinforcement learning from human feedback with multiple preference oracles for trading off performance with safety and fairness constraints.
Unsupervised neural network for 4D Flow MRI velocity field enhancement and phase wrapping correction using divergence-free parameterization.
Physical reservoir computing using Lead Zirconate Titanate for digit classification.
Alignment metrics for comparing neural network representations operating in superposition.
Diversity-aware reverse KL divergence method improving LLM distillation with large capacity mismatches.
Analysis of neural collapse dynamics identifying critical feature norm threshold for convergence.
MAC-Attention acceleration technique for LLM long-context decoding that preserves attention computation fidelity without compression.
Hierarchical flow matching framework for computationally efficient graph generation with reduced complexity.
Knowledge-Data ML framework integrating numeric data with knowledge for model construction.
Apprenticeship learning from imperfect demonstrations with evolving rewards in e-learning contexts.
Research on shuffling strategies for stochastic gradient descent optimization with convergence analysis.
Reinforcement learning framework for autonomous solver selection in chemical kinetics integration.
Agent system using RL to select optimal deep generative models for tabular data synthesis.
Generative framework for subsurface velocity model synthesis using proxy posterior estimation.
Conditional decoding strategy (CASA) for improving safety alignment in multimodal LLMs against cross-modal attacks.
AI safety research on vulnerabilities in autonomous agents with filesystem/email access via circuit analysis.
XGBoost model for startup founder success prediction using engineered features from career data.
Method for encoding graph structure into LLMs via graph pooling tokens for Graph Question Answering tasks.
Deep learning surrogate optimization for production control in stress-sensitive oil reservoirs.
Reinforcement learning approach for behavioral support in Type 1 Diabetes management and insulin dosing.
Gradient-based data valuation for curriculum learning in game-theoretic motion planning using TracIn scoring.
Study showing deep networks assign higher density to simpler out-of-distribution data than in-distribution test data.
Tuning-free GNN prompting framework for cross-graph adaptation without task-specific parameter updates.
Membership inference attack on LLMs via gradient-induced feature drift to detect training data exposure.
Distributed optimization algorithm for Byzantine-resilient gradient tracking with probabilistic edge dropout.
Lagrangian Descriptors framework for evaluating neural network models of Hamiltonian dynamics.