Split and Conquer Partial Deepfake Speech
Split-and-conquer framework for detecting manipulated speech regions via boundary detection and segment-level classification.
Split-and-conquer framework for detecting manipulated speech regions via boundary detection and segment-level classification.
GPU-accelerated solver combining randomized subspace embedding and Nesterov acceleration for large-scale portfolio optimization.
Learning from synthetic data using provenance-based input gradient guidance to teach models discriminative regions.
Inversion-free stochastic natural gradient method for optimization on Riemannian manifolds with implicit parameter constraints.
Fredholm Integral Neural Operators framework for learning non-expansive integral operators with universal approximation proofs.
SkillRT: Compiler-inspired system treating LLM agent skills as code for consistent, efficient execution across diverse platforms.
Theoretical characterization of Gaussian universality breakdown in high-dimensional convex empirical risk minimization under non-Gaussian data.
Analysis of why discrete tokenization limits Vision-Language-Action model scaling; introduces Compression Gap information-theoretic principle.
First transferable learned membership inference attack on fine-tuned LLMs using unlimited labeled data from fine-tuning process itself.
PR3DICTR: PyTorch/MONAI-based open framework for 3D medical image classification and standardized deep learning model training.
Tsetlin Machine-based intrusion detection system for IoMT networks addressing cybersecurity vulnerabilities in medical device systems.
LLM framework for causal graph discovery using breadth-first search queries, reducing complexity from quadratic to linear query requirements.
Methods for training constrained regression trees incorporating domain-specific output constraints using mixed-integer programming and other approaches.
Integration of neural networks into combinatorial optimization for NP-hard problems, learning heuristics and optimality scores via graph convolutional networks.
Amortized inference framework training single model to predict causal mechanisms across multiple datasets for out-of-distribution generalization.
Method for detecting unauthorized training data in one-step distilled diffusion models using distributional statistics instead of memorization detection.
Zero-shot concept bottleneck models enabling interpretable predictions without target task training by leveraging pre-trained vision-language models.
Mathematical framework unifying diffusion and flow-based generative models as denoising Markov processes with rigorous theoretical foundation.
Integration of linear temporal logic specifications into RL using differentiable simulation for safe, correct-by-construction controller synthesis.
Noise-robust exploration method for RL using learning progress monitoring to escape unlearnable noise sources with improved sample efficiency.
ARMOR: one-shot post-training pruning algorithm for LLMs achieving 2:4 semi-structured sparsity with minimal performance degradation for efficient deployment.
Theoretical analysis of convergence guarantees for decentralized stochastic gradient descent with high-probability bounds and reduced assumptions.
Extension of Forward-Forward algorithm to reinforcement learning with action-conditioned Q-functions, replacing backpropagation with local learning.
f-INE framework for stable influence estimation under training randomness, addressing instability in sample-level impact estimation for data curation.
Dataset distillation method leveraging diffusion models as priors to synthesize compact, representative datasets with improved diversity and generalization.
Deep learning approach for antimicrobial peptide design using semi-supervised latent Bayesian optimization with improved interpretability.
MusicRFM framework enabling fine-grained control over pre-trained music generation models by steering internal activations via Recursive Feature Machines.
Bayesian parameter inference method for complex stochastic simulators using differentiable approaches to reduce simulation costs in high-dimensional spaces.
Goal-driven reward signals from pretrained video diffusion models for reinforcement learning agent training.
Distillation-based continual learning with classifier-proximal plugins addressing stability-plasticity tradeoff.
Method to robustify activation sparsity in LLMs by addressing representational instability during inference acceleration.
Machine learning approach for early chronic kidney disease screening in low-resource settings using explainable models.
Scalable multi-concept unlearning in text-to-image diffusion models addressing weight conflicts and collateral damage.
Analysis of extreme variance in certified robustness verification across neural network model seeds.
Textual Equilibrium Propagation for optimizing compound AI systems with multiple modules in long-horizon agentic workflows.
Early classification of time series in non-stationary environments with uncertain and time-varying decision costs.
ChronoSpike: Adaptive spiking graph neural network for dynamic graph representation learning with event-driven efficiency.
Unified training-serving system combining RL with adaptive speculative decoding for accelerated LLM inference.
Infusion: Framework using influence functions to craft training data perturbations that induce targeted model behavior changes.
Uncertainty quantification for machine learning interatomic potentials using evidential deep learning.
Geometric analysis of optimization dynamics in transformers trained on modular arithmetic revealing low-dimensional subspaces.
Study on early-warning signals of grokking via loss-landscape geometry on SCAN and Dyck-1 benchmarks.
CeRA: Parameter-efficient fine-tuning method extending LoRA with non-linear capacity expansion via gating and dropout.
Physics-informed neural operators for solving PDEs with improved generalization beyond training distributions.
SafeSci: Framework for evaluating safety of large language models in scientific domains with comprehensive benchmarks.
CRISP: Method for teaching LLMs to reason more concisely via self-distillation with 'be concise' conditioning.
Stock market prediction using Node Transformer and BERT sentiment analysis for financial forecasting.
WinDiNet uses pretrained video diffusion model as differentiable physics simulator for urban wind flow prediction, replacing expensive CFD simulations.
λ-GELU parameterized gating function enabling controlled ReLU conversion while maintaining smooth activation properties for deployment.
ERPO method for token-level credit assignment in LLM reasoning models, addressing entropy collapse in GRPO through information heterogeneity.