AI-generated summaries

Today's ML research,
without the noise.

Daily summaries of the latest machine learning papers from arXiv, processed every 8 hours.

24 Papers today
8h Update frequency
7 Days of history
Borrowed Geometry: Computational Reuse of Frozen Text-Pretrained Transformer Weights Across Modalities
Abay Bektursun
Multimodal Robotics Large Language Models
  • Frozen text-pretrained transformer weights can be reused across different modalities.
  • Significant performance gains were observed in robotic manipulation tasks using the frozen weights.
  • The study identifies specific attention heads that are crucial for task performance across modalities.
  • The methodology demonstrates that pretrained weights can serve as a general computational substrate.
Read more
What Physics do Data-Driven MoCap-to-Radar Models Learn?
Kevin Chen, Kenneth W. Parker, Anish Arora
Interpretability
  • Introduction of a physics-based interpretability framework for MoCap-to-radar models.
  • Development of two metrics to evaluate physical consistency without requiring ground truth radar data.
  • Demonstration that low reconstruction error does not guarantee physical consistency.
  • Identification of temporal attention as a critical factor for transformer models in learning physics.
Read more
A Dirac-Frenkel-Onsager principle: Instantaneous residual minimization with gauge momentum for nonlinear parametrizations of PDE solutions
Matteo Raviola, Benjamin Peherstorfer
Optimization Theory
  • Introduces the Dirac-Frenkel-Onsager principle to address non-uniqueness in parameter dynamics.
  • Utilizes a history variable interpreted as momentum to promote smooth parameter evolution.
  • Maintains the instantaneous residual minimization property of the Dirac-Frenkel principle.
  • Demonstrates increased robustness in challenging computational regimes.
Read more
Caracal: Causal Architecture via Spectral Mixing
Bingzheng Gan, Tianyi Zhang, Yusu Li, Jing Huang, Wei Shi, Yangkai Ding, Tao Yu
NLP Large Language Models Efficient ML
  • Caracal introduces a Multi-Head Fourier (MHF) module that replaces traditional attention mechanisms, achieving O(L log L) complexity.
  • The architecture employs frequency-domain causal masking to enforce autoregressive capabilities, addressing a critical barrier for Fourier-based models.
  • Caracal eliminates the need for explicit positional encodings by leveraging the inherent properties of the Fourier Transform.
  • The model demonstrates competitive performance against Transformer and SSM baselines while maintaining portability and ease of implementation.
Read more
AlphaInventory: Evolving White-Box Inventory Policies via Large Language Models with Deployment Guarantees
Chenyu Huang, Jianghao Lin, Zhengyang Tang, Bo Jiang, Ruoqing Jiang, Benyou Wang, Lai Wei
Large Language Models Reinforcement Learning Optimization
  • AlphaInventory integrates large language models with reinforcement learning for evolving inventory policies.
  • The framework provides statistical safety guarantees for policy deployment in dynamic environments.
  • A theoretical interface connects training, inference, and deployment, characterizing performance gaps.
  • Empirical results show AlphaInventory outperforms traditional and deep learning inventory strategies.
Read more
Learning physically grounded traffic accident reconstruction from public accident reports
Yanchen Guan, Haicheng Liao, Chengyue Wang, Zhenning Li
Multimodal
  • Introduces a multimodal learning framework for traffic accident reconstruction from public reports.
  • Develops the CISS-REC dataset with 6,217 real-world accident cases.
  • Achieves improved reconstruction fidelity, including accident point accuracy and collision consistency.
  • Demonstrates the potential of using public accident reports as scalable data for traffic safety analysis.
Read more
Uniform-Correct Policy Optimization: Breaking RLVR's Indifference to Diversity
Anamika Lochab, Bolian Li, Ruqi Zhang
Reinforcement Learning Optimization Large Language Models
  • Identifies a fundamental limitation in RLVR objectives regarding probability mass distribution among correct solutions.
  • Proposes Uniform-Correct Policy Optimization (UCPO) to address diversity collapse in RLVR.
  • Theoretically characterizes the optimal policy structure using robustness and entropy-regularized optimality criteria.
  • Demonstrates significant improvements in Pass@K and diversity metrics without compromising Pass@1 performance.
Read more
Deep Kernel Learning for Stratifying Glaucoma Trajectories
Bruce Rushing, Angela Danquah, Alireza Namazi, Arjun Dirghangi, Heman Shakeri
Time Series NLP Multimodal
  • Introduction of a hybrid architecture combining clinical-BERT embeddings with a DKL algorithm for predicting glaucoma patient trajectories.
  • Identification of three clinically distinct patient subgroups, emphasizing the importance of trajectory over current disease severity.
  • Demonstration of superior performance compared to standard time-series forecasting methods.
  • Provision of calibrated uncertainty estimates to support clinical decision-making.
Read more
CRADIPOR: Crash Dispersion Predictor
Edgar Chaillou, Sebastian Rodriguez, Yves Tourbier, Francisco Chinesta
Theory Optimization
  • CRADIPOR addresses the issue of numerical dispersion in automotive crash simulations.
  • The proposed method combines Rank Reduction Autoencoder with supervised classification.
  • RRAE outperforms Random Forest in identifying regions sensitive to numerical dispersion.
  • Wavelet-based and slope-based signal representations are most effective for classification.
Read more
AdaMeZO: Adam-style Zeroth-Order Optimizer for LLM Fine-tuning Without Maintaining the Moments
Zhijie Cai, Haolong Chen, Guangxu Zhu
Large Language Models Optimization Efficient ML
  • AdaMeZO leverages Adam-style moment estimates without storing them, significantly reducing memory requirements.
  • The optimizer achieves faster convergence compared to MeZO, requiring up to 70% fewer forward passes.
  • Theoretical convergence bounds are established, showing AdaMeZO's effectiveness in non-convex scenarios.
  • Extensive experiments validate AdaMeZO's superior performance across multiple LLM architectures.
Read more
Aitchison Embeddings for Learning Compositional Graph Representations
Nikolaos Nakis, Chrysoula Kosma, Panagiotis Promponas, Michail Chatzianastasis, Giannis Nikolentzos
Graph Learning
  • Introduction of Aitchison Compositional Graph embeddings (AICoG) for graph representation learning.
  • Use of isometric log-ratio (ILR) coordinates to preserve Aitchison distances and enable optimization in Euclidean space.
  • Enhanced interpretability through a geometric notion of roles based on compositional latent space.
  • Subcompositional coherence allows for principled component restriction and analysis of archetype influence.
Read more
Unlearning Offline Stochastic Multi-Armed Bandits
Zichun Ye, Runqi Wang, Xuchuang Wang, Xutong Liu, Shuai Li, Mohammad Hajiesmaili
Reinforcement Learning Theory Efficient ML
  • First study of unlearning in offline stochastic multi-armed bandits.
  • Formalization of privacy constraints and utility measurement in decision-making.
  • Development of adaptive algorithms combining Gaussian mechanism and rollback.
  • Theoretical performance guarantees and lower bounds established.
Read more
The Power of Order: Fooling LLMs with Adversarial Table Permutations
Xinshuai Dong, Haifeng Chen, Xuyuan Liu, Shengyu Chen, Haoyu Wang, Shaoan Xie, Kun Zhang, Zhengzhang Chen
NLP Large Language Models Optimization
  • LLMs exhibit significant vulnerability to the layout of tabular data, leading to inconsistent outputs.
  • The Adversarial Table Permutation (ATP) attack is introduced as a method to systematically identify harmful permutations.
  • Extensive experiments show that ATP can degrade the performance of a wide range of LLMs, regardless of their size or architecture.
  • The study reveals a fundamental weakness in the 'linearize-then-prompt' paradigm used in current TQA tasks.
Read more
RunAgent: Interpreting Natural-Language Plans with Constraint-Guided Execution
Arunabh Srivastava, Mohammad A. (Amir) Khojastepour, Srimat Chakradhar, Sennur Ulukus
NLP Large Language Models Interpretability
  • RunAgent integrates natural language processing with structured programming constructs for reliable task execution.
  • The platform autonomously generates and validates constraints for each step of the workflow.
  • RunAgent supports dynamic selection of execution strategies, enhancing flexibility and accuracy.
  • Human-in-the-loop features allow for user specification and feedback, improving the auditing process.
Read more
Possibilistic Predictive Uncertainty for Deep Learning
Yao Ni, Jeremie Houssineau, Yew Soon Ong, Piotr Koniusz
Theory Efficient ML
  • Introduction of a new framework for epistemic uncertainty modeling using possibility theory.
  • Derivation of a tractable implementation with closed-form solutions for efficient computation.
  • Demonstration of competitive performance against leading uncertainty quantification methods across diverse datasets.
Read more
Temporal Data Requirement for Predicting Unplanned Hospital Readmissions
Ramin Mohammadi, Vahab vahdat, Sarthak Jain, Amir T. Namin, Ramya Palacholla, Sagar Kamarthi
Time Series NLP Efficient ML
  • Shorter observation windows (3-6 months) are optimal for predicting readmissions using clinical notes.
  • Structured data models improve with longer observation periods but plateau after 12 months.
  • The study challenges the assumption that more historical data always leads to better predictive performance.
  • Different data types (structured vs. unstructured) require distinct approaches for optimal model performance.
Read more
High-Probability Convergence in Decentralized Stochastic Optimization with Gradient Tracking
Aleksandar Armacki, Haoyuan Cai, Ali H. Sayed
Optimization Federated Learning Theory
  • Introduces GT-DSGD, a decentralized optimization method incorporating gradient tracking.
  • Achieves high-probability convergence under relaxed assumptions compared to traditional DSGD.
  • Establishes optimal HP convergence rates for non-convex and Polyak-Lojasiewicz costs.
  • Provides the first HP guarantees for decentralized optimization methods with bias-correction.
Read more
Technical Report: Activation Residual Hessian Quantization (ARHQ) for Low-Bit LLM Quantization
YiFeng Wang, Zhun Sun, Keisuke Sakaguchi
Large Language Models Efficient ML Optimization
  • ARHQ effectively mitigates error propagation in low-bit quantization of LLMs.
  • The method isolates error-sensitive weight directions using a residual Hessian approach.
  • Experimental results show improved SNR and reasoning performance in quantized models.
  • ARHQ adapts to specific quantization hardware and conditions, enhancing robustness.
Read more
Revealing graph bandits for maximizing local influence
Alexandra Carpentier, Michal Valko
Graph Learning Theory Efficient ML
  • Introduces a graph bandit framework that does not require prior knowledge of the graph structure.
  • Proposes BARE, a bandit strategy that learns to identify influential nodes through limited feedback.
  • Establishes a regret guarantee that scales with the detectable dimension rather than the number of nodes.
  • Demonstrates the practical applicability of the method in marketing scenarios involving social networks.
Read more
Generating Statistical Charts with Validation-Driven LLM Workflows
Pavlin G. Poličar, Andraž Pevcin, Blaž Zupan
Multimodal Large Language Models
  • Introduces a structured workflow for generating statistical charts from tabular data.
  • Emphasizes the importance of rendered-output validation to improve chart readability and semantic accuracy.
  • Retains comprehensive multimodal representations for each generated chart, including code, context, and Q&A pairs.
  • Demonstrates the workflow's effectiveness through the generation of 1,500 charts and evaluation of multimodal LLMs.
Read more
Odysseus: Scaling VLMs to 100+ Turn Decision-Making in Games via Reinforcement Learning
Chengshuai Shi, Wenzhe Li, Xinran Liang, Yizhou Lu, Wenjia Yang, Ruirong Feng, Seth Karten, Ziran Yang, Zihan Ding, Gabriel Sarch, Danqi Chen, Karthik Narasimhan, Chi Jin
Reinforcement Learning Multimodal Robotics
  • Introduces Odysseus, a framework for training VLMs in long-horizon decision-making tasks.
  • Proposes a lightweight turn-level critic in PPO to improve training stability and efficiency.
  • Demonstrates the advantages of pretrained VLMs in providing strong action priors.
  • Achieves at least 3× improvement in game progress compared to frontier models.
Read more
Scale-Aware Adversarial Analysis: A Diagnostic for Generative AI in Multiscale Complex Systems
Mengke Zhao, Guang-Xing Li, Duo Xu, Keping Qiu
Generative Models Theory Interpretability
  • Introduction of a scale-aware diagnostic framework using Constrained Diffusion Decomposition (CDD).
  • Demonstration of the limitations of traditional XAI methods in capturing physical causality.
  • Evaluation of Denoising Diffusion Probabilistic Models (DDPM) under physical perturbations.
  • Establishment of CDD-based scale-space continuity as a criterion for physically consistent deep learning.
Read more
Towards Robust and Scalable Density-based Clustering via Graph Propagation
Yingtao Zheng, Hugo Phibbs, Ninh Pham
Graph Learning Efficient ML Theory
  • CluProp reimagines density-based clustering as a graph propagation process, improving robustness and scalability.
  • The framework is agnostic to distance metrics and employs a deterministic algorithm for efficient neighborhood identification.
  • CluProp significantly outperforms existing clustering methods in both accuracy and runtime on large-scale datasets.
  • DANE algorithm allows for effective label propagation from local density peaks, enhancing clustering performance in heterogeneous data.
Read more
Distance metric learning for conditional anomaly detection
Michal Valko, Milos Hauskrecht
Theory Optimization Time Series
  • Conditional anomaly detection allows for context-specific identification of anomalies.
  • Instance-based approaches optimize predictive models for individual data instances.
  • Standard distance metrics may be inadequate for anomaly detection tasks.
  • Metric-learning methods (NCA and RCA) improve anomaly detection performance.
Read more