Temporal Dropout Risk in Learning Analytics: A Harmonized Survival Benchmark Across Dynamic and Early-Window Representations
Survival-oriented benchmark for temporal student dropout risk modeling using Open University Learning Analytics Dataset.
Survival-oriented benchmark for temporal student dropout risk modeling using Open University Learning Analytics Dataset.
Temporal survival modeling framework for predicting student dropout using LMS engagement data and administrative records.
Re-examines capacity gap in chain-of-thought distillation, finding student models often outperform teacher distillation baselines.
HTNav framework for aerial vision-and-language navigation combining visual perception with language instructions in urban environments.
HM-Bench benchmark evaluates multimodal LLMs on hyperspectral remote sensing image understanding tasks.
Analysis of causal inference methods applied to graph representation learning and their limitations with graph-structured data.
ADRUwAMS deep learning model with attention mechanisms for automated brain tumor glioma segmentation in medical imaging.
Ge2mS-T improves energy efficiency in Spiking Vision Transformers through multi-dimensional grouping and optimized training methods.
UDG dataset with 300K samples for training defect/anomaly generation models with improved generalization across defect categories.
RAG systems should optimize for utility (task completion) rather than topical relevance when retrieving documents for LLMs.
MuTSE: Human-in-the-loop evaluator tool for systematically comparing LLM text simplification outputs across different prompting strategies and architectures.
WOMBET: Framework for reinforcement learning that generates and transfers experience data between source and target robotic tasks for sample efficiency.
Aligned Agents, Biased Swarm: Empirical study measuring how multi-agent system topologies and feedback loops amplify bias in emergent behaviors.
Litmus ReAgent: Benchmark and agentic system for evaluating multilingual LLM performance prediction across 1,500 questions spanning six tasks and five evidence scenarios.
Neighbourhood Transformer: Graph neural network architecture using switchable attention to handle heterophilic graph learning where dissimilar nodes are frequently connected.
PerMix-RLVR: Training method for aligning LLM personas with reward models while preserving output diversity, avoiding inference-time computation overhead.
PinpointQA dataset and benchmark for evaluating small object localization and spatial reasoning in video MLLMs.
ASTRA: adaptive semantic tree reasoning architecture for LLM-based complex table question answering.
Survey and construction of linguistically-informed representations for English as a second/foreign language.
Named entity identification and anonymization system for cybercrime datasets using speech-to-text and image processing.
Regime-conditional retrieval with transferable router for two-hop question answering with theoretical foundations.
Noise-aware in-context learning approach to mitigate hallucinations in auditory large language models.
ImageProtector prevents multi-modal LLMs from analyzing images via visual prompt injection attacks.
Vision-language models for image geolocation with structured geographic reasoning and autonomous self-evolution.
CONDESION-BENCH evaluates LLM decision-making with compositional action spaces and conditional feasibility constraints.
U-Cast: simple probabilistic weather forecasting using standard U-Net architecture achieving frontier performance.
Watt Counts: open-access energy consumption benchmark for LLM inference across 50 models and 10 GPU architectures.
PDYffusion combines diffusion models with physics-informed dynamics for long-horizon spatiotemporal prediction.
Vision-Language-Action models for autonomous driving combining perception, reasoning, and temporal dynamics modeling.
Frequency-enhanced diffusion models for zero-shot skeleton action recognition in computer vision.
NyayaMind framework for transparent legal reasoning and judgment prediction in Indian courts using LLMs.
Method integrating graph-based embeddings into event sequence models for improved user prediction on digital platforms.
DeepGuard improves secure code generation by LLMs through multi-layer semantic aggregation to mitigate vulnerable patterns.
CLIP-Inspector detects backdoor attacks in prompt-tuned vision-language models through out-of-distribution trigger inversion.
Research on detecting covert misaligned AI behavior in real-world settings using open-source intelligence methods.
TensorHub introduces Reference-Oriented Storage for efficient weight transfer in LLM reinforcement learning across heterogeneous computational resources.
PS-TTS method for phonetic synchronization in automated dubbing, addressing duration and lip-sync challenges in AI-based video translation.
Interactive ASR system with human-like interaction and semantic coherence evaluation, replacing WER metric with agent-based correction mechanisms.
EquiformerV3: SE(3)-equivariant graph attention Transformer for 3D atomistic modeling, improving efficiency, expressivity, and physical consistency.
CORA framework for risk-controlled GUI automation agents using conformal prediction to provide formally verified, user-tunable safety guarantees for VLM-powered mobile automation.
LLM-based agents for scaffolding diagnostic reasoning in educational settings, combining scenario-based learning with learning analytics and personalized support.
Dataset for personality-shaped emotional responses to text events, addressing limitations of LLM role-playing and personality illusion in affective computing.
Theoretical analysis of generalization and scaling laws for Mixture-of-Experts Transformers, separating active capacity from routing combinatorics with covering-number bounds.
Symbolic-Neural Consistency Audit framework extracting and formalizing LLM self-stated safety policies.
Vision transformer application predicting chemotherapy response in ovarian cancer from preoperative CT scans.
GNN-based deep reinforcement learning scheduler for cloud workflow DAG assignment minimizing time and energy.
GRM gradient-ratio masking attack on audio LLMs balancing jailbreak success with utility preservation.
Computational model of Von Economo neurons implementing biological speed-accuracy tradeoff in decision-making.
Neural distribution prior method for LiDAR out-of-distribution detection in autonomous driving.
Statistical analysis of I-Ching King Wen sequence showing no improvements to neural network training.