AI-generated summaries

Today's ML research,
without the noise.

Daily summaries of the latest machine learning papers from arXiv, processed every 8 hours.

24 Papers today
8h Update frequency
7 Days of history
AIM-DDI: A Model-Agnostic Multimodal Integration Module for Drug-Drug Interaction Prediction
Yerin Park, Sangseon Lee
Multimodal
  • AIM-DDI is a model-agnostic integration module for DDI prediction.
  • It maps heterogeneous drug modalities into a shared latent space as tokens.
  • The module improves prediction performance, particularly in unseen-drug scenarios.
  • AIM-DDI shows significant relative improvements in accuracy and recall metrics.
Read more
Crys-JEPA: Accelerating Crystal Discovery via Embedding Screening and Generative Refinement
Nian Liu, Nikita Kazeev, Stephen Gregory Dale, Artem Maevskiy, Yuwei Zeng, Ryoji Kubo, Pengru Huang, Thomas Laurent, Yann LeCun, Kostya S. Novoselov, Xavier Bresson
Generative Models
  • Identification of a stability-novelty trade-off in crystal generation.
  • Development of Crys-JEPA, an energy-aware latent surrogate for stability evaluation.
  • Introduction of a screening-and-refinement pipeline to improve generative model performance.
  • Significant performance improvements over baseline models on crystal generation metrics.
Read more
R2R2: Robust Representation for Intensive Experience Reuse via Redundancy Reduction in Self-Predictive Learning
Sanghyeob Song, Donghyeok Lee, Jinsik Kim, Sungroh Yoon
Reinforcement Learning Robotics Theory
  • R2R2 addresses representation-level instability in SPL under high UTD regimes.
  • The method avoids zero-centering to preserve critical global dynamics information.
  • R2R2 improves the performance of TD7 by approximately 22% at a UTD ratio of 20.
  • SimbaV2-SPL, enhanced with R2R2, sets a new state-of-the-art in continuous control benchmarks.
Read more
GeoViSTA: Geospatial Vision-Tabular Transformer for Multimodal Environment Representation
Yuhao Liu, Sadeer Al-Kindi, Ashok Veeraraghavan, Guha Balakrishnan
Multimodal
  • GeoViSTA integrates gridded imagery and tabular socioeconomic data for comprehensive geospatial analysis.
  • The model employs bilateral cross-attention and a geography-aware attention mechanism for effective feature fusion.
  • Training is conducted using a self-supervised joint masked autoencoding objective, requiring no labeled data.
  • GeoViSTA outperforms traditional models in predicting health-related and environmental outcomes.
Read more
Action-Inspired Generative Models
Eshwar R. A., Debnath Pal
Generative Models
  • Introduction of Action-Inspired Generative Models (AGMs) to improve generative model training.
  • Utilization of a lightweight learned scalar potential, VΟ•, to score bridge samples and modulate drift objectives.
  • Significant improvements in generation quality by selectively penalizing uninformative transport paths.
  • VΟ• adds negligible overhead to training and no cost during inference, making it a practical enhancement.
Read more
PreFT: Prefill-only finetuning for efficient inference
Andrew Lanpouthakoun, Aryaman Arora, Zhengxuan Wu, Dhruv Pai, Ben Keigwin, Dan Jurafsky, Christopher Potts
NLP Large Language Models Efficient ML
  • PreFT optimizes inference efficiency by applying adapters only during the prefill phase.
  • The approach significantly increases throughput, achieving up to 1.9Γ— the throughput of traditional PEFTs.
  • PreFTs maintain competitive performance with traditional PEFTs, especially in reinforcement learning tasks.
  • The implementation is available on the vLLM inference engine, facilitating practical applications.
Read more
RxEval: A Prescription-Level Benchmark for Evaluating LLM Medication Recommendation
Shuhao Chen, Weisen Jiang, Changmiao Wang, Xiaoqing Wu, Xuanren Shi, Yu Zhang, James T. Kwok
Large Language Models NLP
  • RxEval shifts the evaluation of medication recommendation from admission-level to prescription-level, capturing the dynamic nature of clinical decision-making.
  • The benchmark includes 1,547 MCQs based on real patient data, enhancing the realism of the evaluation process.
  • Evaluation results show that even state-of-the-art LLMs perform poorly on RxEval, indicating significant challenges in automated medication recommendation.
  • Common errors identified include overlooking patient information and failing to derive clinical conclusions, suggesting areas for model improvement.
Read more
NeuroAtlas: Benchmarking Foundation Models for Clinical EEG and Brain-Computer Interfaces
Konstantinos Kontras, Trui Osselaer, Stylianos G. Mouslech, Angeliki-Ilektra Karaiskou, Guido Gagliardi, Thomas Strypsteen, Mohammad Hossein Badiei, Anku Rani, Maarten Vanmarcke, Miguel Bhagubai, Chanakya Ekbote, Jaedong Hwang, Christos Chatzichristos, Paul Pu Liang, Maarten De Vos
Time Series
  • NeuroAtlas is the largest EEG benchmark with 42 datasets and ~260k hours of EEG data.
  • EEG-specific foundation models do not consistently outperform generic time-series models.
  • Standard metrics are inadequate for assessing clinical utility; bespoke evaluation measures are necessary.
  • Model performance varies significantly across datasets within the same domain.
Read more
Comparative Evaluation of Machine Learning Approaches for Minority-Class Financial Distress Prediction Under Class Imbalance Constraints
Karan Sehgal, Khawar Naveed Bhatti
Interpretability
  • The study emphasizes the importance of sensitivity to minority-class predictions in financial distress scenarios.
  • A comprehensive machine learning workflow is proposed, integrating preprocessing, imbalance mitigation, and explainability.
  • Multiple machine learning models are evaluated, showcasing the effectiveness of ensemble methods in handling class imbalance.
  • SHAP explainability methods are applied to enhance interpretability and governance in financial distress predictions.
Read more
MetaMoE: Diversity-Aware Proxy Selection for Privacy-Preserving Mixture-of-Experts Unification
Weisen Jiang, Shuhao Chen, Sinno Jialin Pan
Federated Learning Computer Vision NLP
  • MetaMoE enables the unification of independently trained experts without sharing private data, ensuring privacy.
  • The framework employs diversity-aware proxy selection using a relevance-weighted DPP to enhance the representation of client domains.
  • A proxy-aligned expert training strategy is introduced, aligning expert behavior with proxy data for better coordination.
  • The context-aware router improves expert assignment across diverse input types.
Read more
LiSA: Lifelong Safety Adaptation via Conservative Policy Induction
Minbeom Kim, Lesly Miculicich, Bhavana Dalvi Mishra, Mihir Parmar, Phillip Wallis, Bharath Chandrasekhar, Kyomin Jung, Tomas Pfister, Long T. Le
NLP Large Language Models Reinforcement Learning
  • LiSA formulates lifelong guardrail adaptation for AI agents using sparse, noisy user feedback.
  • The framework incorporates broad policy abstraction, conflict-aware local rules, and confidence-gated memory reuse.
  • Empirical results show LiSA consistently outperforms strong baselines and maintains robustness against noisy inputs.
  • LiSA enhances boundary-sensitive decision-making and pushes the latency-performance frontier beyond traditional scaling methods.
Read more
XFP: Quality-Targeted Adaptive Codebook Quantization with Sparse Outlier Separation for LLM Inference
Thomas Witt
Large Language Models Efficient ML Optimization
  • XFP allows operators to specify quality thresholds instead of bit-widths, enhancing flexibility in quantization.
  • The quantizer automatically determines codebook size and outlier handling, simplifying the quantization process.
  • XFP achieves effective bits as low as 3.4 while maintaining high throughput for large models.
  • Two codebook storage modes provide options for balancing precision and memory usage.
Read more
ROAD: Adaptive Data Mixing for Offline-to-Online Reinforcement Learning via Bi-Level Optimization
Letian Yang, Xu Liu, Yiqiang Lu, Jian Liu, Weiqiang Wang, Shuai Li
Reinforcement Learning Optimization Theory
  • Identifies an objective misalignment problem in existing O2O RL approaches.
  • Proposes a bi-level optimization framework for adaptive data mixing.
  • Utilizes a multi-armed bandit mechanism for real-time optimization of data mixing ratios.
  • Achieves superior stability and performance compared to static and heuristic-based methods.
Read more
Compositional Sparsity as an Inductive Bias for Neural Architecture Design
Hongyu Lin, Antonio Briola, Yuanrong Wang, Tomaso Aste
Theory Efficient ML Interpretability
  • Introduces a novel architecture combining IFNs and HNNs to exploit compositional sparsity.
  • HNNs are significantly sparser than standard DNNs, requiring less hyperparameter tuning.
  • Empirical results show HNNs outperform dense architectures in both synthetic and real-world datasets.
  • The approach provides a stable and interpretable framework for high-dimensional learning.
Read more
Turning Stale Gradients into Stable Gradients: Coherent Coordinate Descent with Implicit Landscape Smoothing for Lightweight Zeroth-Order Optimization
Chen Liang, Xiatao Sun, Qian Wang, Daniel Rakita
Optimization Efficient ML Theory
  • CoCD is a deterministic and memory-efficient ZO optimizer that utilizes stale gradients effectively.
  • Theoretical grounding connects CoCD to established optimization methods, proving its equivalence to BCCD with warm starts.
  • Larger finite-difference step sizes can improve convergence stability by smoothing the optimization landscape.
  • Empirical results show CoCD significantly outperforms existing methods in terms of sample efficiency and accuracy.
Read more
Optimal Pattern Detection Tree for Symbolic Rule-Based Classification
Young-Chae Hong, Yangho Chen
Interpretability Optimization
  • Introduction of the Optimal Pattern Detection Tree (OPDT) for symbolic rule-based classification.
  • Utilization of mixed-integer programming to discover a single optimal pattern in data.
  • Incorporation of Branching Structure Constraints (BSC) to encode domain knowledge and compliance requirements.
  • Demonstrated optimality guarantees and effective performance on real-world datasets.
Read more
A Novel Schur-Decomposition-Based Weight Projection Method for Stable State-Space Neural-Network Architectures
Sergio Vanegas, Lasse Lensu, Fredy Ruiz
Theory Efficient ML Time Series
  • Introduction of a Schur-stable state-matrix weight-projection scheme.
  • Alternative stable state-matrix parameterization for improved computational efficiency.
  • Demonstrated performance on synthetic and real-world datasets.
  • Achieves comparable accuracy and convergence rates to state-of-the-art methods.
Read more
Distributionally Robust Multi-Task Reinforcement Learning via Adaptive Task Sampling
Nicholas E. Corrado, Wenyuan Huang, Josiah P. Hanna
Reinforcement Learning Optimization Robotics
  • DRATS addresses imbalanced data allocation in MTRL by prioritizing tasks with the largest return gap.
  • The algorithm is derived from a minimax optimization framework, focusing on minimizing the worst-case return gap.
  • Empirical results show DRATS achieves higher data efficiency and better performance on harder tasks compared to existing methods.
  • DRATS can be integrated with existing multi-task learning architectures, enhancing their effectiveness.
Read more
TFGN: Task-Free, Replay-Free Continual Pre-Training Without Catastrophic Forgetting at LLM Scale
Anurup Ganguli
Large Language Models NLP Theory
  • TFGN enables continual pre-training without task labels or data replay.
  • Achieves minimal backward transfer and high retention rates across diverse text domains.
  • Demonstrates positive cross-domain forward transfer, enhancing model performance.
  • Introduces extensions for autonomous continual learning and effective forward-pass behavior.
Read more
Watch your neighbors: Training statistically accurate chaotic systems with local phase space information
Joon-Hyuk Ko, Andrus Giraldo, Deok-Sun Lee
Theory Time Series Optimization
  • Introduces a framework that bridges the gap between Jacobian accuracy and long-term statistical behavior in chaotic systems.
  • Utilizes local coverings of chaotic attractors to analyze dynamics and improve model training.
  • Demonstrates significant improvements in Jacobian accuracy even in the presence of noise.
  • Provides a practical solution for training surrogate models without requiring ground-truth dynamics.
Read more
Novel Dynamic Batch-Sensitive Adam Optimiser for Vehicular Accident Injury Severity Prediction
Daniel Asare Kyei, Alimatu Saadia-Yussiff, Maame G. Asante-Mensah, Abdul Lateef-Yussiff, Charles Roland Haruna, Derry Emmanuel
Optimization Time Series
  • Introduction of DBS-Adam, an optimiser that adapts learning rates based on batch difficulty.
  • Integration of DBS-Adam with Bi-LSTM networks for predicting injury severity in vehicular accidents.
  • Demonstrated significant improvements in model performance metrics compared to traditional optimisers.
  • Achieved a test accuracy of 95.22% and improved precision, recall, and F1-score.
Read more
Lang2MLIP: End-to-End Language-to-Machine Learning Interatomic Potential Development with Autonomous Agentic Workflows
Wenwen Li, Yuki Orimo, Nontawat Charoenphakdee
NLP Large Language Models Optimization
  • Lang2MLIP reduces the need for domain expertise in MLIP development by using natural language inputs.
  • The framework employs a multi-agent system to manage the workflow dynamically, allowing for self-correction.
  • Evaluation on a solid electrolyte interphase system shows the effectiveness of the approach in adapting to complex materials.
  • Lang2MLIP represents a shift from fixed pipelines to a more flexible, decision-based model for MLIP development.
Read more
Collaborative Yet Personalized Policy Training: Single-Timescale Federated Actor-Critic
Leo Muxing Wang, Pengkun Yang, Lili Su
Reinforcement Learning Federated Learning Robotics
  • Introduces a federated actor-critic framework that supports personalized policy training.
  • Establishes finite-time convergence rates for critic error and policy gradient norms.
  • Demonstrates linear speedup with respect to the number of agents in heterogeneous environments.
  • Develops new perturbation analysis techniques for projected subspace updates.
Read more
Peng's Q(Ξ») for Conservative Value Estimation in Offline Reinforcement Learning
Byeongchan Kim, Min-hwan Oh
Reinforcement Learning Theory
  • CPQL is the first multi-step Q-learning algorithm for model-free offline RL.
  • The method adapts the PQL operator for conservative value estimation without requiring additional models.
  • Theoretical analyses confirm that CPQL achieves performance greater than or equal to the behavior policy.
  • Extensive experiments show CPQL significantly outperforms existing offline single-step RL algorithms.
Read more