Quantization Meets dLLMs: A Systematic Study of Post-training Quantization for Diffusion LLMs
Systematic study of post-training quantization methods for diffusion LLMs to enable edge device deployment, comparing compression techniques.
Systematic study of post-training quantization methods for diffusion LLMs to enable edge device deployment, comparing compression techniques.
UTRL: reinforcement learning framework training LLMs to generate high-quality unit tests automatically, addressing test generation challenges.
Research evaluating Law-Following AI framework for embedding legal compliance in advanced AI agents, analyzing legal personhood constructs and technical feasibility.
Trajectory-based paradigm for efficient 3D point cloud tracking in robotics and autonomous systems.
Reinforcement learning approach for radiology report generation using FactScore-based rewards with reduced data requirements.
Framework evaluating robustness of Vision-Language-Action models under real-world physical variations for robotic tasks.
Slovak parliamentary speech corpus with 66M words and fine-tuned ASR models for low-resource language recognition.
Data-efficient ASR personalization using phoneme-level uncertainty scoring and variational inference. Guides fine-tuning for non-normative speech recognition.
Variational low-rank adaptation method for personalizing speech recognition on impaired speech using foundation models. Addresses acoustic variability and data scarcity.
Autoregressive method for autonomous driving combining HD map construction with persistent traffic rule awareness across extended driving sequences.
Matched-compute study evaluating synthetic data interventions for in-context learning in language models. Tests mechanism-targeted pretraining effects.
Method for reducing LLM agent inference costs through trajectory reduction. Addresses token cost efficiency in multi-turn agent systems for software engineering.
Continuous-time reinforcement learning theory with deterministic policy gradients for continuous state and action spaces.
Deep reinforcement learning optimization method using eigenspectrum analysis and condition numbers to improve sample efficiency and critic network performance.
Technique reducing LLM reasoning model overthinking through decoupled rewards and curriculum scheduling. Addresses excessive token generation without performance gain.
Variational flow matching approach for vector-quantized image generation combining categorical supervision with continuous transport dynamics.
Diffusion model method for diverse text-to-image generation via contrastive noise optimization. Addresses mode collapse in text-guided image synthesis.
Reinforcement learning framework (SFPO) for improving LLM reasoning by decomposing policy optimization into slow and fast components to reduce training instability.
Information-theoretic framework measuring higher-order structure and emergence in multi-agent LLM systems. Tests for dynamical emergence in agent coordination.
Model-native technique to explain LLM hallucinations using layer-wise semantic maps. Traces concept flow through residual streams via unembedding.
Method for detecting psychological stress from speech using temporal modeling with cross-attention. Treats stress as dynamically evolving phenomenon.
Comprehensive overview of Ultralytics YOLO object detection model evolution from YOLOv5 to YOLO26. Details architectural improvements and benchmarking.
Transformer-based neural network decoder for quantum error correction. Applies deep learning to quantum computing reliability.
Theoretical analysis of knowledge distillation in neural networks from a functional perspective. Decouples compression from architecture reduction.
Deep learning approach using diffusion models to solve optimal power flow problems in electrical grids. Captures multi-valued solution mappings.
Benchmark and methods for editing knowledge in audio-language models without retraining. First work on auditory attribute knowledge editing.
Machine learning method for recognizing daily activities from sensor data using contrastive learning. Addresses ambient assisted living applications.
Justitia scheduling algorithm for fair and efficient execution of task-parallel LLM agents on shared GPUs. Resource scheduling optimization for agent systems.
StreamingTOM training-free token compression for streaming video understanding. Efficiency optimization for video vision-language models.
GlobalRAG reinforcement learning approach for multi-hop question answering with improved global reasoning. RAG system enhancement via RL for better query planning.
Gaze-guided object detection framework for egocentric videos using vision transformers. Computer vision with attention mechanisms.
Annotation scheme for collaborative dialogue corpus capturing speaker perspectives in misunderstandings. Linguistics and NLP dataset annotation work.
Zero-shot robotic grasp detection using vision-language models without training data or retraining. Robotics application leveraging VLMs.
Semi-supervised intent detection framework with active learning correction for voice dialog agents. LLM application for dialog system improvement.
Vision-language model for unified gaze understanding combining detection, target, and object recognition. Multimodal model for attention estimation.
Study of multi-agent LLM cooperation on math problems with analysis of adversarial robustness. Evaluates agent collaboration and vulnerability to perturbations.
Privacy-preserving explainable AI for IoT applications using SHAP entropy regularization. Focuses on interpretability and privacy in edge devices.
Training-free stabilizer (PAS) fixing temporal inconsistency in video LLMs caused by rotary position embedding ripples. Technical improvement for video understanding.
Decoupled action expert for vision-language-action models using diffusion/flow-matching for manipulation policies. Computer vision and robotics research.
Physical adversarial attack against stereo depth estimation in autonomous driving perception. Security research without direct relevance to stated interests.
Fast inference acceleration for masked auto-regressive diffusion models enabling practical reinforcement learning. Optimization technique for generative models in RL contexts.
Pipeline for speaker-attributed civic simulation using LLMs from ASR transcripts. LLM application for multi-agent deliberation modeling.
Concept bottleneck models for explainable visual anomaly detection with semantic interpretability. Computer vision research with interpretability focus.
MapReduce LoRA and RaTE methods for multi-preference optimization in generative models using RLHF. Advances alignment of LLMs through parameter-efficient fine-tuning techniques.
Deep learning model for generating site-specific earthquake ground motion from accelerometer records. Domain-specific application without relevance to stated interests.
Review of uncertainty quantification and data efficiency methods for AI systems in robotics, telecommunications, and healthcare. Theoretical ML research with practical applications.
Method for composing visual concepts from images and videos using prompt token binding. Computer vision research without clear AI agent or LLM application focus.
Combines large language models with transformer encoders for financial news classification with limited labeled training data.
Rough set theory applied to explain results of spectral graph clustering algorithms for text document analysis.
Neuroscience study of cortical neuron mechanisms underlying long-term working memory through spike-timing precision analysis.