Unified Graph Prompt Learning via Low-Rank Graph Message Prompting
Unified graph prompt learning framework using low-rank message prompting for efficiently fine-tuning pre-trained graph neural networks across multiple graph components.
Unified graph prompt learning framework using low-rank message prompting for efficiently fine-tuning pre-trained graph neural networks across multiple graph components.
Framework for predicting antibody-antigen binding affinity using listwise ranking and positive-unlabeled learning to address label sparsity in therapeutic design.
Streaming approximate nearest neighbor index for high-dimensional vectors inspired by mycelium growth patterns, using adaptive topology with edge decay and traffic-driven reinforcement.
Spatio-temporal forecasting method using sheaf diffusion to model heterogeneous local interactions in systems, reformulating prediction as information flow over locally structured spaces.
Federated learning approach using personalized submodels aligned with client computational budgets, addressing structural diversity and representational adaptation across heterogeneous devices.
Modular neural architecture for learning three-valued logic end-to-end across mathematical domains without external symbolic solvers.
Memory-Enhanced Dynamic reward Shaping (MEDS) framework for reinforcement learning that reduces failure pattern recurrence in LLM training.
Post-hoc conformal selection method using e-variables for adaptive false discovery rate control in hypothesis testing.
Theoretical framework for discrete diffusion models on graphs using free-energy gradient flows and Markov chain methods.
Framework for detecting and recovering from sensor attacks in cyber-physical systems using Bayesian inference and active probing.
Framework for training deep learning models on non-differentiable scientific metrics using knowledge-guided learning and Minkowski Image Loss.
Research on optimizing reinforcement learning with verifiable rewards for LLMs using low-rank trajectory modeling to reduce computational training overhead.
Dynamic mixture-of-experts architecture for graph neural networks that adapts expert computation per node based on difficulty.
Analysis of policy-based tariff interventions on global supply chain networks amid geopolitical tensions.
Diffusion language model for goal-directed molecular generation satisfying heterogeneous constraints like protein compatibility and drug properties.
Comparison of KV-cache compression strategies for transformers showing quantization outperforms rank reduction across multiple model sizes.
Analysis of per-sample forgetting dynamics during fine-tuning of image classifiers, examining architecture dependence of retention patterns.
Continuous-time flow models trained with adversarial objectives using learned discriminators instead of fixed MSE criteria.
Evaluation framework and benchmark for assessing time-series foundation models and forecasting approaches.
Semi-supervised reinforcement learning approach using knowledge-enhanced data synthesis to improve medical reasoning in LLMs.
Python package providing bioacoustic deep learning models and evaluation tools for analyzing passive acoustic monitoring data.
Study of multi-class linear classification in transformers using feature and label permutation equivariance to extract interpretable algorithms.
Analytical formalism decomposing Hessian matrices in neural networks with DAG architectures into inter-layer blocks.
Framework for autonomous mechanistic reasoning in virtual cells using LLMs, representing biological reasoning as mechanistic action graphs.
Methodology to mitigate shortcut learning and demographic bias in deep neural networks using geometric a priori approaches.
Model-free reinforcement learning system for autonomous crystal alignment using visual information without domain knowledge of crystallography.
ClawGUI framework for training, evaluating, and deploying GUI agents that interact with software through visual interfaces, with online RL and evaluation infrastructure.
Mechanistic analysis of looped reasoning language models examining internal dynamics and latent state evolution compared to standard feedforward models.
Uses reinforcement learning on physics simulators to train models solving Physics Olympiad problems, addressing lack of large-scale physics QA datasets for reasoning models.
SHANG++: Accelerated stochastic gradient descent methods robust to multiplicative noise in gradient updates.
LABBench2: Improved benchmark for evaluating AI systems and agents on biology research tasks with real-world capabilities.
VTC: DNN compilation method using virtual tensors to eliminate data movement in neural network workloads including LLMs.
Pipeline and best practices for log analysis in AI systems to understand model behaviors, with code examples in Inspect framework.
User study evaluating effectiveness of interval-based counterfactual explanations for improving understanding and trust in black-box models.
Benchmark measuring humanization and anti-detection capabilities of mobile GUI agents against platform countermeasures.
Demonstrating LLMs can generate UI interfaces and content together with proper prompting and tool integration.
Object-Oriented Programmatic World Modeling (OOWM) for embodied reasoning and planning in robotic tasks using LLMs.
MobiFlow: Benchmark for mobile agents using trajectory fusion for real-world GUI task evaluation without system-level APIs.
Architecture for maintaining persistent identity in AI agents through multi-anchor memory to prevent catastrophic forgetting.
Spatial Competence Benchmark (SCBench) evaluating large models on spatial reasoning, environment representation, and planning tasks.
ECHO: Speculative decoding optimization for LLM inference in high-concurrency serving with sparse gating.
Benchmark evaluating LLMs as text-only controllers for exploration and navigation in gridworlds under partial observability.
Method for self-calibrating LLMs at test-time through discriminative distillation to reduce overconfidence without labeled data.
Dataset and annotation study of COVID-19 vaccination regret experiences from YouTube comments.
ML framework for predicting 5G network downlink performance using measurements from commercial smartphones.
Machine learning models to assess media literacy skills and identify disinformation among students.
Deep learning model using TCN and attention-based LSTM for predicting stock repurchases in Chinese financial markets.
Study of backdoor security vulnerabilities in flow-matching Vision-Language-Action models used for robotics, exploiting vector field dynamics.
NeuroPath system for motor imagery decoding from EEG signals for brain-computer interfaces in prosthetics and rehabilitation.
Lightweight speech activity-based approach for real-time voicemail detection in telephony using tree ensemble classification.