AI-generated summaries

Today's ML research,
without the noise.

Daily summaries of the latest machine learning papers from arXiv, processed every 8 hours.

24 Papers today
8h Update frequency
7 Days of history
QDSB: Quantized Diffusion Schrödinger Bridges
Tobias Fuchs, Florian Kalinke, Nadja Klein
Generative Models Efficient ML Optimization
  • Introduction of QDSB for efficient generative modeling from unpaired samples.
  • Anchor-based quantization method improves computational efficiency of coupling.
  • Stability analysis provides geometric principles for anchor selection.
  • Empirical results show QDSB matches existing methods in sample quality with reduced time.
Read more
QuIDE: Mastering the Quantized Intelligence Trade-off via Active Optimization
Xiantao Jiang
Efficient ML Optimization Computer Vision
  • Introduction of the Intelligence Index (I) for unified evaluation of quantized neural networks.
  • QuIDE framework offers a standardized protocol for measuring model efficiency across various architectures.
  • Empirical findings reveal a task-dependent Pareto Knee, with optimal bit-widths varying by task complexity.
  • The accuracy-gated variant (I′) effectively identifies non-viable quantization configurations.
Read more
Improving the Performance and Learning Stability of Parallelizable RNNs Designed for Ultra-Low Power Applications
Julien Brandoit, Arthur Fyon, Damien Ernst, Guillaume Drion
Efficient ML Time Series Theory
  • Identified gradient blocking during BMRU updates as a key limitation.
  • Introduced CMRU and αCMRU as novel parallelizable RNN cells with persistent memory.
  • Demonstrated improved convergence stability and reduced initialization sensitivity.
  • CMRU and αCMRU outperform traditional RNNs on tasks requiring long-term retention.
Read more
ADMM-Q: An Improved Hessian-based Weight Quantizer for Post-Training Quantization of Large Language Models
Ryan Lucas, Mehdi Makni, Xiang Meng, Adam Deng, Rahul Mazumder
Large Language Models Optimization Efficient ML
  • ADMM-Q formulates layer-wise weight quantization as a constrained optimization problem, improving upon traditional methods.
  • The algorithm employs a joint optimization approach rather than a greedy column-wise method, enhancing model utility.
  • ADMM-Q shows significant performance improvements over GPTQ, particularly in low-bit quantization scenarios.
  • The method is modular and compatible with existing quantization techniques, ensuring ease of integration.
Read more
RT-Transformer: The Transformer Block as a Spherical State Estimator
Peter Racioppo
Theory
  • Introduction of the Radial–Tangential SDE (RT-SDE) for structured stochastic modeling of noise in Transformers.
  • Attention is reinterpreted as a precision-weighted estimator of latent directions on a hypersphere.
  • Unified derivation of attention, residual connections, and normalization as components of a single filtering update.
  • Proposed architectural modifications enhance the Transformer by incorporating magnitude-dependent precision and normalization.
Read more
No More, No Less: Task Alignment in Terminal Agents
Sina Mavali, David Pape, Jonathan Evertz, Samira Abedini, Devansh Srivastav, Thorsten Eisenhofer, Sahar Abdelnabi, Lea Schönherr
NLP Large Language Models Theory
  • Introduces the Task Alignment Benchmark (TAB) to evaluate task alignment in terminal agents.
  • Defines task alignment as the selective use of environmental information, distinguishing between cue utilization and distraction resistance.
  • Demonstrates a gap between task capability and task alignment in existing terminal agents.
  • Shows that suppressing distractor execution can also suppress necessary cues for task completion.
Read more
KAN-CL: Per-Knot Importance Regularization for Continual Learning with Kolmogorov-Arnold Networks
Minjong Cheon
Computer Vision Theory Efficient ML
  • KAN-CL utilizes per-knot importance regularization to address catastrophic forgetting in continual learning.
  • The framework combines a KAN classification head with a convolutional backbone, enhancing feature extraction while localizing task-specific parameters.
  • Significant reductions in forgetting (88% and 93%) were achieved on standard benchmarks, while maintaining competitive accuracy.
  • Theoretical insights from NTK analysis support the effectiveness of the proposed method in preventing forgetting.
Read more
TRACE: Temporal Routing with Autoregressive Cross-channel Experts for EEG Representation Learning
Fan Ma, Qier An, Peng Chen, Lingfei Qian, Xiang Lan, Mingyang Jiang, Zhiling Gu, Xenophon Papademetris, Hua Xu
Time Series
  • TRACE introduces an autoregressive EEG pre-training framework that predicts future EEG patches from causal context.
  • The TR-MoE block combines spatial-temporal attention with a cross-channel routing mechanism to maintain coherence across channels.
  • TRACE supports heterogeneous pre-training across various EEG datasets without requiring a uniform electrode layout.
  • The framework achieves leading results on multiple EEG benchmarks, demonstrating its effectiveness in both seen and unseen domains.
Read more
Targeted Neuron Modulation via Contrastive Pair Search
Sam Herring, Jake Naviasky, Karan Malhotra
NLP Large Language Models Interpretability
  • CNA identifies the 0.1% of MLP neurons crucial for distinguishing harmful from benign prompts.
  • Neuron-level ablation reduces refusal rates by over 50% while preserving output coherence.
  • Refusal mechanisms in instruction-tuned models are targetable and sparse, unlike base models.
  • Results are consistent across different architectures and model sizes, indicating robustness.
Read more
Incentivizing Truthfulness and Collaborative Fairness in Bayesian Learning
Rachael Hwee Ling Sim, Jue Fan, Xiao Tian, Xinyi Xu, Patrick Jaillet, Bryan Kian Hsiang Low
Theory Federated Learning
  • Introduces a mechanism that ensures both collaborative fairness and truthfulness in data sharing.
  • Combines semivalues with a truthful data valuation function based on an unknown validation set.
  • Proves the existence of a truthful equilibrium where sources maximize rewards through honest data submission.
  • Addresses the limitations of existing methods that either ensure fairness or truthfulness, but not both.
Read more
From Generic Correlation to Input-Specific Credit in On-Policy Self Distillation
Guobin Shen, Lei Huang, Xiang Cheng, Chenxiao Zhao, Jindong Li, Dongcheng Zhao, Xing Yu
NLP Large Language Models Reinforcement Learning
  • Self-distillation rewards in language models can be interpreted as Bayesian filtering increments measuring pointwise mutual information.
  • There exists an input-generic bias in self-distillation rewards that can dilute the effectiveness of credit assignment.
  • The proposed CREDIT method effectively isolates input-specific contributions, improving performance on various benchmarks.
  • CREDIT enhances learning efficiency with negligible additional computational cost.
Read more
Causal Fairness for Survival Analysis
Drago Plecko
Theory Interpretability Time Series
  • Introduces a causal framework for fairness in survival analysis, addressing temporal disparities.
  • Develops a non-parametric four-step methodology for causal pathway decomposition.
  • Proves the Causal Reduction Theorem to facilitate the identification of group disparities.
  • Applies the framework to analyze racial disparities in ICU outcomes, illustrating its practical utility.
Read more
SOAR: Scale Optimization for Accurate Reconstruction in NVFP4 Quantization
Chengzhu Bao, Xianglong Yan, Zhiteng Li, Guangshuo Qin, Guanghua Yu, Yulun Zhang
NLP Large Language Models Efficient ML
  • SOAR improves NVFP4 quantization accuracy through innovative scale optimization techniques.
  • Closed-form Joint Scale Optimization (CJSO) allows for simultaneous optimization of global and block-wise scales.
  • Decoupled Scale Search (DSS) mitigates precision loss by separating quantization and dequantization scales.
  • Extensive experiments show SOAR outperforms existing methods while maintaining the same memory footprint.
Read more
Deep Minds and Shallow Probes
Su Hyeong Lee, Risi Kondor
Theory
  • Probing methods should be stable under representation symmetries to avoid artifacts from arbitrary coordinate choices.
  • A unique hierarchy of shallow probes is established, with linear probes as the degree-1 member and higher-order probes introduced systematically.
  • The concept of probe-visible quotients is introduced for effective cross-model transfer, focusing on directions visible to the probes rather than the full hidden state.
  • Experiments show that degree-2 probes significantly improve performance over linear probes in specific tasks.
Read more
FedOUI: OUI-Guided Client Weighting for Federated Aggregation
Alberto Fernández-Hernández, Jose I. Mestre, Cristian Pérez-Corral, Manuel F. Dolz, Jose Duato, Enrique S. Quintana-Ortí
Federated Learning
  • FedOUI introduces a new aggregation rule based on the Overfitting-Underfitting Indicator (OUI).
  • The method improves client weighting by considering the internal activation structure of client models.
  • Empirical results show significant improvements in aggregation quality under strong data heterogeneity.
  • FedOUI remains lightweight and interpretable, making it suitable for practical federated learning applications.
Read more
A Proof-of-Concept Simulation-Driven Digital Twin Framework for Decision-Aware Diabetes Modeling
Zarrin Monirzadeh
Time Series
  • Introduction of a modular digital twin architecture for multiple diabetes types.
  • Transition from correlation-based prediction to decision-aware modeling.
  • Reproducible proof-of-concept evaluation using an open dataset.
  • Discussion of practical deployment considerations including interpretability and safety.
Read more
On the Approximation Complexity of Matrix Product Operator Born Machines
Chao Li, Zerui Tao, Yuchen Cong, Juan Xu, Qibin Zhao
Theory Generative Models Efficient ML
  • Proved that KL approximation for MPO-BMs is NP-hard in the continuous setting.
  • Identified conditions under which MPO-BMs can achieve efficient approximation with polynomial bond dimension.
  • Demonstrated that polynomially many score queries are sufficient for estimating the induced Hamiltonian.
  • Established a connection between score-based variational inference and the ground-state problem in physics.
Read more
Persona-Conditioned Adversarial Prompting: Multi-Identity Red-Teaming for Adversarial Discovery and Mitigation
Cristian Morasso, Anisa Halimi, Muhammad Zaid Hameed, Douglas Leith
Large Language Models NLP
  • PCAP conditions adversarial searches on diverse attacker personas to explore realistic attack vectors.
  • Empirical evaluation shows a significant increase in attack success rates and prompt diversity.
  • Fine-tuning on PCAP-generated data dramatically enhances model robustness with minimal false positives.
  • The approach provides a practical pipeline for automated vulnerability discovery and mitigation.
Read more
Efficient Adjoint Matching for Fine-tuning Diffusion Models
Jeongwoo Shin, Dongsoo Shin, Joonseok Lee, Jaewoong Choi, Jaemoo Choi
Generative Models Optimization Efficient ML
  • EAM significantly improves training efficiency by reformulating the SOC problem.
  • The method eliminates the need for backward adjoint simulation, reducing computational costs.
  • EAM converges up to 4× faster than traditional Adjoint Matching while maintaining or exceeding performance metrics.
  • The approach leverages a linear base drift to facilitate efficient trajectory sampling.
Read more
Physics-Informed Teacher-Student Ensemble Learning for Traffic State Estimation with a Varying Speed Limit Scenario
Archie J. Huang, Dongdong Wang, Shaurya Agarwal, Mohamed Abdel-Aty, Md Mahmudul Islam, Muhammad Shahbaz
Theory Optimization Time Series
  • Integration of teacher-student ensemble learning with PIDL for TSE under VSL scenarios.
  • Teacher models encode local traffic physics while the student model selects appropriate estimations.
  • Demonstrated superior performance in TSE compared to traditional methods.
  • Addresses the challenges of varying traffic characteristics due to dynamic speed limits.
Read more
A Unified Graph Language Model for Multi-Domain Multi-Task Graph Alignment Instruction Tuning
Haibo Chen, Xin Wang, Jiaheng Chao, Ling Feng, Wenwu Zhu
Graph Learning Large Language Models NLP
  • UniGraphLM is the first model to integrate a multi-domain, multi-task GNN encoder with LLMs for unified graph token generation.
  • The proposed graph-text pair pretraining strategy enhances the alignment of GNN representations with textual semantics.
  • A curriculum alignment tuning strategy is introduced to adaptively manage varying alignment difficulties across diverse graph data.
  • Extensive experiments validate the superiority of UniGraphLM over existing GLM baselines in multiple domains and tasks.
Read more
Block-R1: Rethinking the Role of Block Size in Multi-domain Reinforcement Learning for Diffusion Large Language Models
Yan Jiang, Ruihong Qiu, Zi Huang
Reinforcement Learning Large Language Models NLP
  • Introduces the concept of domain block size conflict in multi-domain RL for dLLMs.
  • Develops the Block-R1-41K dataset with optimal block sizes for individual samples.
  • Establishes Block-R1 as a benchmark for cross-domain RL post-training.
  • Proposes a sample-level block-conditioned training method for improved policy updates.
Read more
Hierarchical Multi-Scale Graph Neural Networks: Scalable Heterophilous Learning with Oversmoothing and Oversquashing Mitigation
Md Sazzad Hossen, Avimanyu Sahoo
Graph Learning
  • Introduction of a label-free adaptive signed affinity to prevent sign cancellation in graph learning.
  • Development of the first framework that constructs an orthonormal, multi-scale, sparse spectral basis in near-linear time.
  • Empirical evidence showing that conventional GNNs suffer from hub domination, oversmoothing, and oversquashing, which HMH effectively mitigates.
  • Achievement of state-of-the-art accuracy on both node and graph classification tasks while ensuring linear scalability.
Read more
Stochastic Minimum-Cost Reach-Avoid Reinforcement Learning
Jingduo Pan, Taoran Wu, Yiling Xue, Bai Xue
Reinforcement Learning Robotics Optimization
  • Introduction of Reach-Avoid Probability Certificates (RAPCs) for enforcing probabilistic reach-avoid constraints.
  • Development of a contraction-based Bellman formulation that integrates safety and cost optimization.
  • Proposal of RAPCPO, a reinforcement learning algorithm that converges to locally optimal policies under probabilistic constraints.
  • Demonstration of improved cost efficiency and high satisfaction rates in stochastic environments through experiments.
Read more