D5P4: Partition Determinantal Point Process for Diversity in Parallel Discrete Diffusion Decoding
D5P4 framework applies determinantal point processes to discrete diffusion decoding for diverse parallel text generation.
D5P4 framework applies determinantal point processes to discrete diffusion decoding for diverse parallel text generation.
Algorithm for generalized symmetric matrix factorization with exactness properties and non-Lipschitz optimization.
Method for splitting pretrained language models into specialized domain-specific models using continued pretraining strategies.
Multi-agent framework for grounding vision-language navigation using probabilistic reasoning about spatial relations and metric constraints.
Evaluates State Space Models as vision encoders for Vision-Language Models, comparing SSM backbones to transformer-based alternatives.
DreamPartGen generates semantically grounded 3D objects with part-level decomposition using text-to-3D diffusion methods.
DriveTok proposes efficient 3D tokenization for multi-view driving scenes to improve autonomous driving systems and world models.
Nemotron-Cascade 2: 30B open-weight MoE LLM with strong reasoning and agentic capabilities, achieving IMO Gold Medal performance.
Method for designing adaptive noise schedules in diffusion models for image and video generation using spectral guidance.
NavTrust benchmark evaluates trustworthiness of embodied navigation agents under real-world corruptions in Vision-Language Navigation and Object-Goal Navigation tasks.
Establishes improved learning rates for stochastic gradient descent and Nesterov accelerated gradient with generalization performance guarantees.
Chat Incremental Pattern Constructor extracts ordered token-transition rules from text for interpretable machine learning rule extraction.
Optimization methods for inverse classification problems including counterfactual explanations and adversarial examples using logistic and softmax classifiers.
CADGL uses context-aware deep graph learning for predicting drug-drug interactions with improved generalization and robustness.
μLO derives Maximal Update Parametrization for learned optimizers to improve meta-generalization across network widths and unseen tasks.
Flow matching approach with large-scale synthetic dataset for solving inverse ellipsometry problem of reconstructing optical film properties.
ODE-constrained generative model for synthesizing realistic 12-lead ECG training data to address scarcity of labeled medical recordings.
Cliqueformer uses structured transformers for model-based optimization in design problems like protein engineering via offline learning.
VOGP algorithm using Gaussian process bandits for black-box vector optimization with incomplete order relations and Pareto optimality guarantees.
Theoretical analysis showing shallow nonlinear networks learn linearly separable features with polynomial width scaling relative to data dimension.
Methods to achieve real-world efficiency gains from token filtering in LLM training through improved sparsity and adaptive filtering strategies.
Survey of Part-Prototype Models for explainable AI, examining interpretability mechanisms and competitive limitations versus alternative approaches.
Two neural architectures for precipitation nowcasting integrating weather station data and radar measurements for improved forecast skill.
OPUS-VFL addresses privacy-utility tradeoffs and incentive mechanisms in Vertical Federated Learning with heterogeneous client resources.
Causal intervention framework for interpreting Variational Autoencoders mechanistically, addressing interpretability of generative models.
Shapley Value-based alternating training framework for multimodal fusion that balances dominant and minor modalities.
Statistical framework for fairness testing in algorithmic systems that accounts for sampling error and handles intersectional demographic analysis.
Analysis of communication scheduling in decentralized learning showing benefits of concentrating synchronization in later training stages.
GeoReg uses LLMs with satellite imagery and geospatial data for socio-economic indicator estimation in data-scarce regions via few-shot regression.
Research on Online Convex Optimization algorithms for heavy-tailed gradient distributions, extending beyond finite variance assumptions.
Physics-informed neural network framework predicting fatigue life of steels under nuclear reactor conditions.
Neural network architecture using nonharmonic Fourier series for scientific machine learning applications.
Transformer architecture with dual attention for multivariate time-series anomaly detection using temporal invariants.
Theoretical analysis of parameter norm scaling in overparameterized linear regression and diagonal networks.
Framework evaluating faithfulness of chain-of-thought reasoning in large audio language models for multimodal tasks.
Flow-matching models for 3D point cloud generation using optimal transport and meanflow for single-step inference acceleration.
KAN-based feature selection framework for tabular data via spline-based importance scoring. Specialized ML technique.
Sub-quadratic attention algorithm removing bounded-entry restrictions for LLM inference speedup. Foundational LLM efficiency research.
Quantization technique for vision encoders using prefix registers to handle outliers. Optimization research for multimodal models.
Theoretical reinforcement learning on decision-estimation coefficients for adversarial MDPs. Pure RL theory.
Conditional flow matching for precipitation forecasting. Weather prediction ML, not core AI interests.
Diffusion-Transformer model converting images directly to G-code for 3D printing. Applied ML, domain-specific.
Genetic algorithm for sample reweighting to mitigate ML bias. Fairness-focused, not primary tech interests.
Continual learning research on replay buffer size impact on feature retention vs. classifier forgetting. Specialized ML theory.
Neural ODEs for quantum many-body dynamics simulation. Physics-focused ML, not core AI interests.
Algorithm extraction from Discrete Transformers via symbolic program synthesis. Addresses representation entanglement in interpretability.
EEG foundation model for brain-computer interfaces with biophysical grounding. Neuroscience domain, not AI/tech stack focused.
Research analyzing mechanistic changes when post-training autoregressive models into masked diffusion models. Studies model internals via circuit analysis.
Machine learning for materials science: multimodal models predict dielectric elastomer properties under limited data. Domain-specific ML, not AI-focused.
Unified theoretical framework for model merging explaining effectiveness across heterogeneous fine-tuning hyperparameters with scaling laws.