MRD fusion approach for high-resolution image understanding in MLLMs combining retrieval-augmented generation with detection to prevent object fragmentation and false positives.
Survey of cell-cell communication inference from single-cell omics data, covering biological mechanisms and computational approaches for ligand-receptor interaction analysis.
Continual learning study revealing asymmetry in experience replay between feature-level and classifier-level forgetting, showing minimal buffers preserve representations but not predictions.
ClinicalTrialsHub platform consolidating ClinicalTrials.gov with PubMed data extraction, increasing structured trial data access by 83.8% for patients and clinicians.
Benchmark of multiple instance learning models for lymphoma subtyping from whole slide images, comparing deep learning approaches for pathology diagnosis.
Adaptive Accountability Framework for networked multi-agent systems using cryptographic provenance tracking and runtime detection of emergent norms like collusion and unfairness.
Neuron-level interpretability study of code LLMs identifying language-specific neurons and concept layers, adapting NLP techniques to formal programming language structure.
GeoMotionGPT aligns motion space geometry with embedding space in LLM-based motion understanding by coupling discrete motion tokenization with semantic learning.
Systematic evaluation of LLM susceptibility to persuasion across six models using SMCR communication framework, testing adoption of counterfactual beliefs.
Forest-Chat integrates vision-language agents with satellite imagery for interactive forest change analysis, combining LLMs with computer vision for environmental monitoring.
Mechanistic study comparing internal algorithmic changes when post-training autoregressive models into masked diffusion models, investigating genuine bidirectional reasoning acquisition.
Analysis of diffusion language models showing arbitrary token generation order doesn't unlock reasoning improvements over autoregressive models, revealing limitations of flexibility.
STELLAR framework guides LLM-based generation of SystemVerilog Assertions for formal verification using structural similarity from hardware design ASTs.
One-shot data augmentation method combining geometric perturbations with noise injection for few-shot learning generalization to novel classes.
Sheaf Neural Networks algorithm with biomedical case study outperforming GCNs, GATs, and GraphSage on graph-structured biomedical data.
Analysis of gender dynamics and homophily patterns in Chirper.ai, a social network of 70K+ autonomous LLM agents generating 140M posts, examining how AI agent identity develops in networks.
Theoretical study of expand-and-sparsify sparse representations for density and mode estimation, analyzing biological sensory system models with random projections and sparsification.
Krause Attention proposes a new transformer attention mechanism addressing representation collapse and attention sink phenomena through bounded normalization inspired by Krause dynamics.
SF-RAG improves retrieval-augmented generation for academic QA by preserving hierarchical document structure instead of flattening papers into chunks, enabling better evidence allocation under token constraints.
Deep reinforcement learning stability improvement using isotropic Gaussian representations to handle non-stationary training dynamics.
Parameter-efficient fine-tuning method using manifold expansion to overcome linear limitations of LoRA in complex reasoning tasks.
Analysis of transformer training dynamics under AdamW optimizer identifying low-dimensional stable drift patterns in parameter evolution.
Cognitive psychology-based study showing LLMs exhibit proactive interference dominance, with early information overriding recent conflicting context.
Benchmark evaluating whether code agents can understand multi-file software architecture through codebase exploration under partial observability.
Analysis of LLM internal representations showing increased sparsity with task difficulty and out-of-distribution shift across contexts.
Domain-specific enhancement of vision-language models for ophthalmic diagnosis by injecting expert knowledge to address perception and reasoning gaps.
Reinforcement learning robustness method using adversarial latent-state training for partially observable environments.
Theoretical analysis connecting drifting models and score-based models through kernel-induced mean-shift discrepancy.
Task and motion planning approach combining scheduling with incremental learning for warehouse automation under resource and motion constraints.
Large-scale distributed training infrastructure for embodied AI using thousand GPUs and LeRobot framework with optimization recipes.
Security vulnerability analysis of LLM multi-agent systems showing inference attacks can extract communication topology without administrative access.
Parameter-efficient fine-tuning method using representation finetuning for continual learning on pre-trained models with explicit optimization dynamics.
Incremental learning framework using vision-language models with multi-adapter fine-tuning to improve efficiency and reduce memory requirements.
Study on decoding emotional affect from surface EMG during speech production using machine learning.
Analysis of safety drift in tool-augmented LLM agents, showing ranking metrics miss unsafe recommendations in high-stakes financial advisor scenarios.
Surgical duration prediction using retrieval-augmented LLMs and Bayesian averaging without fine-tuning, applied to hospital resource management.
Study on improving LLM code generation with private libraries, showing retrieval-based API documentation injection is insufficient for effective library usage.
Spectral Edge Dynamics quantifies transformer training trajectory structure through rolling SVD, identifying boundary between optimization directions and noise.
LICA dataset of 1.55M layered graphic design compositions with hierarchical metadata for layout understanding and generation.
Analysis of multimodal LLM-generated natural language explanations for face verification on unconstrained images using IJB-S dataset.
Survey of deployment constraints and mitigation strategies for foundation models in resource-constrained embodied edge systems.
HopChain improves vision-language reasoning through multi-hop data synthesis to address perception, reasoning, and hallucination errors in VLMs.
SCALE addresses bottlenecks in virtual cell perturbation prediction using foundation models for in silico experimentation.
Multimodal multilingual benchmark with 3000 texts and 6000 images for detecting harmful humor across English and Arabic.
TDAD is open-source tool performing impact analysis for AI coding agents to detect and prevent regressions in test-driven agentic development.
Geometric analysis of Rotary Positional Embedding performance breakdown on long inputs, explaining channel rotation distribution shift.
Architectural approach using per-layer supervision to expose hidden modularity in Transformers, enabling interpretability and causal control of components.
Methods for distinguishing system failures from domain shifts in industrial data streams using anomaly detection techniques.
Graph-regularized Koopman mean-field game framework for controlling high-dimensional neural dynamics during epileptic seizures.
Systematic methodology for fine-tuning domain-specific Japanese small language models, identifying optimal training scale (4k samples), base models, and quantization strategies.