MiroThinker-1.7 & H1: Towards Heavy-Duty Research Agents via Verification
MiroThinker-1.7 and H1 research agents with verification for complex long-horizon reasoning and multi-step problem solving.
MiroThinker-1.7 and H1 research agents with verification for complex long-horizon reasoning and multi-step problem solving.
ClawWorm: self-propagating attack demonstrating security vulnerabilities in multi-agent LLM ecosystems like OpenClaw.
Simulation Distillation approach for sim-to-real transfer in robotics; pretrains world models for rapid real-world adaptation.
Framework for computing differentiable geodesics on 3D meshes; enables parallelized Riemannian operators for mesh learning.
Theoretical characterization of partial labels learning feasibility with adaptive nearest neighbors method.
FEEL dataset combining force measurements from piezoresistive gloves with egocentric video for physical action understanding.
Behavioral Foundation Models baseline using regularized latent dynamics prediction for adaptive agent policies.
Theoretical analysis of transformers for knowledge retrieval in LLMs beyond orthogonal embedding assumptions.
Research on dynamic tokenization replacing fixed vocabularies in LLMs; hierarchical autoregressive approach for 70B parameter models.
Constitutional AI research on learning natural language rules automatically for LLM control via multi-agent framework.
Data augmentation framework using pseudo-labeling and unlabeled speech for robust dysarthric speech severity assessment.
Asymmetric pruning technique for vision-language models addressing modality-specific behaviors in text and visual token compression.
LLM-based framework using retrieval augmentation and confidence-based automation for efficient radiology report annotation in clinical NLP.
Power analysis framework for statistical inference on ML-predicted outcomes, addressing sample size determination for prediction-powered inference.
Theoretical analysis of stochastic mirror descent optimization without Lipschitz smoothness using relative smoothness framework.
Feature selection method for distributionally robust learning maintaining reliability across diverse deployment environments with covariate shift.
Attribution upsampling method using redistribution instead of interpolation to prevent corruption of saliency maps in explainable AI.
Parallel in-context learning technique for vision-language models reducing inference latency while maintaining demonstration effectiveness.
Study showing LLM pre-training without learning rate decay improves downstream supervised fine-tuning performance.
Benchmark comparing GAN and Stable Diffusion augmentation strategies for class imbalance correction in animal classification under low-data conditions.
Graph-based multi-agent reinforcement learning for decentralized UAV swarm coordination under partial observability and communication constraints.
Deep Adaptive Design for efficient model-based design of experiments in nonlinear dynamical systems with offline neural network policies.
LLM-based recommender system using review aggregation and multi-factor attention for restaurant recommendations.
Attribution-guided sparse feature steering to mitigate hallucinations in large vision-language models without increasing inference cost.
Deep learning method for discovering error patterns in automotive diagnostic trouble codes and vehicle system fault characterization.
Machine learning framework for predicting drug response in non-small cell lung cancer using genetic and lifestyle data.
YOLO-based deep learning framework for automated wasp identification with explainable AI integration for biodiversity assessment.
1.25B-word corpus for Pashto with reproducible NLP pipeline, deduplication, and quality filtering across 39 sources.
Diffusion policy for robot motion balancing efficiency and legibility in human-robot collaboration through style conditioning.
Reinforcement learning approach training virtual fish to control real fish schools, using 2D screen-displayed agents as alternatives to physical robots.
High-fidelity benchmark with 220 real-world 4K videos for unsupervised physical parameter estimation and governing-equation identification.
Graph theory algorithm for segmentation of detonation cells from 3D pressure traces in detonations research.
Multi-modal adversarial attacks exposing vulnerabilities in image generation model unlearning without full retraining.
Omanic benchmark for step-level evaluation of multi-hop reasoning in LLMs with annotations for diagnosing reasoning failures.
Framework for resource-aware LLM reasoning in embodied robotic agents using reinforcement learning to balance computation and action execution.
Data-driven nonlinearity identification method for mechanical systems using neural network activation functions.
Evaluation of cultural biases in LLMs through author profiling from song lyrics, detecting gender and ethnicity inference in zero-shot settings.
Formal model for selecting statements that find common ground across diverse population preferences using generative AI.
Analysis of conformal factuality robustness in retrieval-augmented generation LLM systems, proposing novel metrics for hallucination evaluation.
Pipeline generating 100K data-generation-ready 3D digital object twins from single images for robotic manipulation simulation.
Analysis of geometric imbalance problem in semi-supervised node classification on class-imbalanced graph data.
Method for detecting fairwashing in black-box algorithmic auditing by identifying compliant surrogate models versus discriminatory production systems.
Decoupled solutions for Linear Model of Co-regionalization multitask Gaussian process model reducing computational complexity.
Factor analysis method for identifying latent psychopathology factors in clinical questionnaire data with improved interpretability.
Heterogeneous graph deep learning model for drug response prediction with interpretability via attention mechanisms.
Research on correcting automatic speech recognition errors using compact seq2seq models trained on real and synthetic ASR error patterns, avoiding LLM latency and hallucination issues.
Deep operator learning for full waveform inversion addressing source generalization by training on diverse seismic source conditions.
TS-Reasoner: domain-specialized LLM agent for multi-step time series reasoning and analysis, integrating language model reasoning with domain-specific computation.
Hypergraph convolutional transformer for QoS prediction handling data sparsity and cold-start issues in service recommendations.
Learning-augmented sketches for frequency estimation in data streams without ground truth labels, improving over traditional memory-constrained methods.