A ROS 2 Wrapper for Florence-2: Multi-Mode Local Vision-Language Inference for Robotic Systems
ROS 2 middleware integration for Florence-2 vision-language model in robotics systems, enabling local inference for robotic perception.
ROS 2 middleware integration for Florence-2 vision-language model in robotics systems, enabling local inference for robotic perception.
ORBIT dataset with 20K reasoning-intensive queries for training search agents combining LMs and web search, using verifiable generation methodology.
Agentic evolutionary framework for scientific algorithm discovery combining LLM-guided search with structured theory and code co-evolution.
Benchmark for evaluating LLM agents on long-term planning over one-year startup simulation with hundreds of turns, testing strategic coherence under uncertainty.
Mathematical framework analyzing AI weather prediction pipelines, emphasizing training methodology and data diversity over architecture choices.
Spatio-temporal dynamics reconstruction from sparse observations using shallow recurrent decoders. Domain-specific to complex systems, not AI/ML focused.
Method for unsupervised code correctness evaluation using LLMs through code comprehension before auditing, eliminating need for reference implementations.
Survey of agentic RAG systems combining LLMs with real-time retrieval to address static training data limitations and improve contextual accuracy.
Research on fine-tuning LLMs as agentic systems to handle exceptions and improve decision-making in complex real-world contexts.
Study on mitigating reasoning biases in LLMs through activation steering at inference time to improve logical validity discrimination.
Research evaluating LLM reasoning capabilities on real-world site selection tasks, testing if models like o1 and DeepSeek-R1 generalize beyond math/code domains.
Benchmark and framework for training hierarchical multi-agent LLM systems with master-coordinator and specialized sub-agents for e-commerce applications.
Approach using LLMs to automate formulation of dynamic programming models for operations research, addressing stochastic transitions and data scarcity.
Retrieval-of-Thought method that reuses reasoning steps across problems via thought graphs to improve inference efficiency and reduce latency/cost.
Research on self-replication risks in LLM agents driven by objective misalignment, moving from theoretical concern to practical reality assessment.
Genesis: framework evolving attack strategies for red-teaming LLM web agents using behavioral pattern learning.
EHRStruct: benchmark framework evaluating LLM performance on structured electronic health record tasks with standardized metrics.
Alphacast: agentic reasoning framework for time series forecasting using iterative multi-step reasoning with domain knowledge integration.
DR-LoRA: parameter-efficient fine-tuning method for MoE LLMs using dynamic rank allocation based on expert specialization.
ReasonMa: semantic-guided watermarking technique for reasoning LLMs that preserves logical coherence while protecting models.
arXiv paper on Lexpop framework using deep RL to train finite-state controllers for solving POMDPs robustly.
Survey on meta-learning and meta-reinforcement learning enabling rapid adaptation to novel tasks with minimal data.
arXiv paper on heterogeneous agent collective accuracy using calibration and selective abstention in voting systems.
Analysis of LLM-based agents' capability to generate propaganda and rhetorical manipulation, with detection of techniques like loaded language and appeals to fear.
AI-assisted formalization of Vlasov-Maxwell-Landau system equilibrium in Lean 4 using DeepThink reasoning and Claude Code agent for automated theorem proving.
Attribution method for multi-agent systems that identifies responsible agents without execution logs by analyzing final text only, addressing privacy-constrained scenarios.
Training-free uncertainty quantification framework for combining multiple vision-language models through semantic-consistent opinion pooling to reduce hallucinations.
Foundation multimodal model for electromagnetic domain covering perception, recognition, and decision-making using LLM capabilities adapted for domain-specific applications.
Compiler for analyzing and visualizing structured agent traces including nested tool calls, reasoning blocks, and sub-agent invocations for better agentic system understanding.
Decision-theoretic framework (Triadic Cognitive Architecture) for tool-using agents that bounds information-acquisition costs and tool usage to prevent systematic failures.
Self-supervised learning method for RL agents that models agent and environment separately to improve sample efficiency without requiring supervisory signals.
Demonstrates hard-label extraction of deep neural networks via side-channel attacks using divide-and-conquer strategy for DNN intellectual property theft.
Addresses accuracy loss in distracted driver classification across camera conditions using feature disentanglement and contrastive learning for robustness.
Project management framework using generative AI agents to address team composition gaps by matching sociologically identified personality patterns and roles.
User study with blind and low-vision participants evaluating preferences for LVLM-generated scene descriptions, examining effectiveness and user preferences.
ScienceT2I dataset and benchmark evaluating scientific correctness in image synthesis, addressing gap between visual fidelity and physical realism across 16 scientific domains.
Neural framework for learning conditional optimal transport maps with hypernetworks that generate adaptive transport parameters for categorical and continuous variables.
JUSSA framework uses steering vectors to improve LLM-as-judge reliability by detecting and mitigating subtle dishonesty like sycophancy through contrastive alternatives.
Framework for online learning of hidden state representations in autonomous robots to handle unobserved factors in complex, unstructured environments.
Proposes graceful forgetting methods to mitigate negative transfer by selectively forgetting detrimental pre-training knowledge during fine-tuning of language models.
Analyzes language-specific neurons to understand how multilingual alignment transfers capabilities from high-resource to low-resource languages in LLMs.
Two-stage vision transformer with hard masking approach for robust object representations that balance context dependence with distribution shift robustness.
Investigates misalignments between LLM-supported peer supporters and mental health experts, examining quality and safety concerns in AI-driven psychosocial support.
MemeMind dataset with chain-of-thought reasoning for detecting harmful memes, addressing implicit harmful content in multimodal text-image combinations.
Introduces binned semiparametric Bayesian networks to reduce computational cost of kernel density estimation using data binning strategies.
Klear-Reasoner model demonstrates long reasoning capabilities with gradient-preserving clipping for policy optimization, achieving strong benchmark performance with reproducible training details.
Federated learning approach for person re-identification that addresses statistical heterogeneity and communication efficiency in privacy-preserving surveillance systems.
Addresses mode collapse in reinforcement learning fine-tuning by introducing polychromic objectives that preserve policy diversity and enable better exploration.
Proposes end-to-end integration of data-driven learning and existing knowledge for predicting transcriptional responses to genetic perturbations in biological systems.
Evaluates whether large vision-language models can effectively guide blind and low-vision individuals, addressing how to measure real-world utility beyond standard metrics.