EvoLen: Evolution-Guided Tokenization for DNA Language Model
EvoLen: evolution-guided tokenization approach for DNA language models. Addresses fundamental tokenization design challenges in biological sequence modeling.
EvoLen: evolution-guided tokenization approach for DNA language models. Addresses fundamental tokenization design challenges in biological sequence modeling.
Experience replay for LLM post-training RL formalizing optimal buffer design as trade-off between sample efficiency and data freshness.
Tensor decomposition method quantifying uncertainty in LLM-based multi-agent systems accounting for communication and role dependencies.
CLOVER framework for multi-agent RL cooperation conditioning value decomposition on realistic wireless communication graphs.
LottaLoRA training paradigm showing frozen random backbones with trained LoRA adapters recover 96-100% performance across diverse tasks.
Adversarial sensor error framework for robust wind turbine fleet control against measurement errors and hacking.
Adaptive simulation experiment framework using pairwise comparisons to optimize LLM policies for operations management tasks.
Prompt optimization method decomposing reward variance into response and prompt variance to identify task amenability to optimization.
Multi-agent actor-critic reinforcement learning for disaster resilience controlling power, communication, and emergency response systems.
Evaluation metrics for SVG generation via element-level structural analysis using leave-one-out evaluation.
4-bit floating-point format (HiFloat4) for efficient language model pre-training on Ascend NPU hardware.
Guidance method for consistency models using joint flow distribution learning to enable classifier-free guidance without separate teacher model.
Training curriculum method for discrete flow-based image generation models to improve one-step sampling stability and quality.
Analysis of LoRA adapter spectral geometry to identify fine-tuning objectives and predict harmful model behavior in language models.
Safety steering mechanism for multimodal LLMs using dictionary-aligned concept control to prevent unsafe outputs without retraining.
Theoretical analysis of finite-sample properties and identifiability bounds for nonlinear Independent Component Analysis algorithms.
Survival analysis benchmark for predicting student dropout in learning analytics using OULAD dataset with dynamic and static representations.
Demonstrates power-law scaling of classification error with number of classes and how chain-of-thought decomposition reduces error through task splitting.
Temporal modeling framework for predicting student dropout using LMS data and logistic regression with counterfactual policy simulation.
Practical analysis of chain-of-thought distillation from students to teachers, revisiting capacity gap assumptions and baseline comparisons.
Conformal prediction framework for transformers providing uncertainty quantification and calibration for trustworthy LLM deployment.
Analysis of causal inference applications in graph representation learning and risks of aggregating graph elements.
Adaptive Thompson sampling for high-dimensional Bayesian optimization addressing sparse candidate point grids.
Using synthetic data for ML-based childhood vaccination prediction in nomadic populations in Kenya.
Dynamic policy optimization bridging SFT and RL for LLMs, addressing bias-variance tradeoff in post-training through adaptive loss weighting.
Empirical study on effectiveness of advanced optimizers for multi-task learning, identifying overlooked factors in optimization approaches.
Analyzes calibration and paraphrase sensitivity in medical vision-language models using predictive entropy and uncertainty quantification.
SeqComm-DFL: Multi-agent coordination via sequential communication and decision-focused learning for value-aware message generation.
WOMBET: World model-based framework for experience transfer in robotics RL, generating and utilizing prior data for sample efficiency.
Hierarchical implicit flow Q-learning for offline goal-conditioned reinforcement learning with improved policy expressiveness.
SentryFuse: Framework for efficient multimodal model compression on edge devices with sensor dropout robustness via zero-shot pruning.
Graph neural network architecture addressing heterophilic graphs using switchable attention mechanism for monophily-aware learning.
Named entity identification and anonymization system for cybercrime datasets on Telegram with speech-to-text transcription.
Solution to NeurIPS 2023 LLM Efficiency Challenge: Fine-tuning LLaMA 70B on single A100 GPU within 24-hour constraint.
U-Cast: Efficient probabilistic weather forecasting model using standard U-Net architecture, simplifying state-of-art approaches.
PDYffusion: Diffusion model for long-horizon spatiotemporal prediction incorporating physics-based constraints and uncertainty quantification.
Proposes PML-MA method for partial multi-label learning using feature-label modal alignment to handle noisy labels.
arXiv paper proposing Temporal Patch Shuffle data augmentation for time series forecasting preserving temporal coherence and improving generalization.
arXiv paper integrating graph-based embeddings into event sequence models for user-item interactions in fraud and recommendation systems.
arXiv paper on GeoPAS geometric probing approach for automated algorithm selection in continuous black-box optimization.
arXiv paper on EquiformerV3, advancing SE(3)-equivariant graph attention Transformers for efficiency, expressivity, and 3D atomistic modeling.
arXiv paper proposing score-driven rating system extending classical Elo rating to accommodate diverse game outcomes and rankings.
arXiv paper on CORA, conformal risk-controlled GUI agents using vision language models with formal safety guarantees for mobile automation.
arXiv paper proposing truncated rectified flow policy for maximum entropy RL enabling one-step multimodal action distribution sampling.
arXiv paper augmenting distillation process dataset with simulations for deep learning-based anomaly detection in chemical batch processes.
arXiv paper developing generalization and scaling theory for Mixture-of-Experts Transformers with covering-number bounds and routing overhead analysis.
arXiv paper on GNN-based deep reinforcement learning scheduler for cloud workflow DAGs optimizing completion time and energy consumption.
arXiv paper analyzing statistical properties of ancient I-Ching King Wen sequence, finding no improvement to neural network training.
arXiv paper proposing DiffHLS framework using GNNs and LLM code embeddings for high-level synthesis quality prediction via differential learning.
arXiv paper investigating LLM pretraining geometry and common minima to improve downstream generalization without changing loss function.