AdvDINO: Domain-Adversarial Self-Supervised Representation Learning for Spatial Proteomics
AdvDINO: domain-adversarial self-supervised learning framework for spatial proteomics to handle batch effects in biomedical imaging.
AdvDINO: domain-adversarial self-supervised learning framework for spatial proteomics to handle batch effects in biomedical imaging.
Analysis of LLM use in newsmaking across 40,000+ articles using AI-text detectors, showing increased GenAI adoption in local and college media.
COXNet: cross-layer fusion network for detecting tiny objects in multimodal RGB-thermal imagery for surveillance and autonomous navigation.
FedKLPR: federated learning approach for person re-identification with KL-guided pruning to reduce communication overhead and handle non-IID data.
Proximal SFT: supervised fine-tuning method using trust-region constraints to prevent capability deterioration when adapting foundation models to new tasks.
FlexiFlow: lifetime-aware design framework for integrated computation in disposable products using flexible electronics with kHz speeds.
LLM-based synthetic training reduces maritime domain model costs 261x by using LLMs as teachers for small language model training.
Diffusion model for audio-driven facial animation using keyframe augmentation and speech feature decomposition.
FS-DFM enables fast long text generation using few-step diffusion language models with parallel position generation.
StyleBench evaluates trade-offs between structured reasoning styles and efficiency/robustness in LLM inference.
Position paper analyzing measurement gaps in reinforcement learning with verifiable rewards for LLMs on structured tasks.
SecureVibeBench evaluates code generation security of LLM-powered code agents against realistic vulnerability scenarios.
Consistency models as plug-and-play priors for solving inverse problems with reduced neural function evaluations.
T-BiGAN framework combining Transformers and BiGAN for unsupervised anomaly detection in power grid synchrophasor data.
Hybrid deep learning system for EEG-based brain-computer interface wheelchair control using motor imagery.
Mathematical framework interpreting Transformers as discretizations of integro-differential equations.
Semantic segmentation combining light field and LiDAR modalities for autonomous driving scene understanding.
LLM-based system for generating standards-aligned math word problems customized to student interests and ability levels.
Protein language models for fitness prediction interpreted as inverse reinforcement learning on evolutionary sequences.
HiPRAG uses hierarchical process rewards to improve agentic RAG efficiency, reducing over-search and under-search behaviors.
Unified framework analyzing sequence models (Transformers, SSMs, gated RNNs) through coefficient dynamics lens.
Survey of inductive reasoning in LLMs, covering particular-to-general thinking patterns and knowledge generalization capabilities.
RAGen framework for generating domain-specific question-answer pairs to adapt RAG systems to specialized applications.
Risk-sensitive abstention in bandit algorithms for high-stakes AI where errors are irreparable without expert guidance.
Post-processing methods for MRI brain image inpainting to handle lesions and tumors in medical imaging analysis.
Multi-hop reasoning over knowledge graphs using multi-view RAG with LLMs, addressing Transformer attention specialization patterns.
Optimization proxies trained to minimize optimality gaps while providing worst-case guarantees for large-scale batch economic dispatch problems.
SimBench provides first standardized benchmark for evaluating how faithfully LLMs simulate human behaviors across diverse tasks and metrics.
AtlasKV enables RAG systems to integrate billion-scale knowledge graphs efficiently in limited VRAM by avoiding expensive external retrieval modules.
Proposes DistDF for time-series forecasting using Wasserstein alignment to handle autocorrelated label sequences better than standard approaches.
Method to automatically extract and explain what features human feedback data encodes when training language models, addressing unpredictability in RLHF approaches.
Analysis of multilingual reasoning gaps in reasoning language models, showing deficits stem from language understanding failures in low-resource languages.
Method for interpreting LLM reasoning by resampling multiple chain-of-thought branches to measure causal influence and underlying computation.
LLM-guided decompilation framework using context to improve re-executability of decompiled binaries for security analysis.
Multimodal diffusion approach for robot learning from expert trajectories, modeling interactions between observations, actions, and rewards.
SynthAgent: Framework for web agent adaptation using synthetic data generation with quality filtering to handle hallucinations and trajectory noise.
GroupRank: Efficient passage reranking paradigm using LLMs with groupwise ranking to balance efficiency and accuracy.
LiveCLKTBench: Benchmark pipeline for reliably measuring cross-lingual knowledge transfer in multilingual LLMs with time-sensitive queries.
Framework for process-centric evaluation of agentic software systems, analyzing execution trajectories and reasoning beyond outcome metrics.
Theoretical framework for sparse dictionary learning in neural networks, analyzing piecewise biconvexity and spurious minima in mechanistic interpretability.
WisPaper: AI agent system for academic paper discovery and organization, addressing semantic search and workflow fragmentation challenges.
Multimodal expert fusion approach for interpretable Alzheimer's disease diagnosis from neuroimaging data.
VPR-AttLLM framework using LLM semantic reasoning to improve geo-localization of crowdsourced flood imagery.
Method for multi-subject image generation with distinction capability, integrating composition and distinction in subject-driven synthesis.
Multimodal RAG system enhanced with knowledge graphs for audio-visual retrieval, extending LLM capabilities to multimodal domains.
Study on imitation learning for autonomous driving, addressing the gap between privileged expert demonstrations and sensor-limited student observations in simulation.
Research on variance-aware tree policies for Monte Carlo Tree Search, improving upon UCB-based methods used in AlphaZero-style algorithms.
CricBench: benchmark for evaluating LLMs on multilingual cricket analytics and domain-specific Text-to-SQL tasks.
Survey of Brazilian K-12 teachers' perceptions on AI in education, examining AI literacy and adoption across 346 educators.
Research questions whether small proxy model training reliably guides data curation decisions for full-scale frontier AI model pretraining.