AI-generated summaries

Today's ML research,
without the noise.

Daily summaries of the latest machine learning papers from arXiv, processed every 8 hours.

66 Papers today
8h Update frequency
7 Days of history
CARE: Covariance-Aware and Rank-Enhanced Decomposition for Enabling Multi-Head Latent Attention
Zhongzhu Zhou, Fengxiang Bie, Ziyan Chen, Zhenyu Zhang, Yibo Yang, Junxiong Wang, Ben Athiwaratkun, Xiaoxia Wu, Shuaiwen Leon Song
Large Language Models Efficient ML
  • CARE enhances the expressivity of attention models without increasing KV-cache costs.
  • The method introduces activation-preserving factorization to align approximations with input activations.
  • Adjusted-rank allocation optimizes the distribution of KV budgets across layers based on their spectral characteristics.
  • CARE outperforms traditional SVD-based approaches, demonstrating significant improvements in perplexity and accuracy.
Read more
Foundations of Schrödinger Bridges for Generative Modeling
Sophia Tang
Generative Models Theory Optimization
  • Schrödinger bridges serve as a unifying framework for various generative modeling techniques.
  • The paper develops both static and dynamic formulations of the Schrödinger bridge problem.
  • A comprehensive toolkit for constructing Schrödinger bridges is provided, facilitating task-specific applications.
  • The framework connects to modern generative modeling approaches, enhancing their theoretical foundation.
Read more
Do Understanding and Generation Fight? A Diagnostic Study of DPO for Unified Multimodal Models
Abinav Rao, Sujan Rachuri
Multimodal Optimization Generative Models
  • DPO cannot improve generation quality in Janus-Pro's VQ-based architecture.
  • Magnitude imbalance between understanding and generation gradients is the primary interference mechanism.
  • Dynamic gradient reweighting can preserve understanding gains in multi-task DPO.
  • The findings are consistent across different model scales (1B and 7B parameters).
Read more
Pathology-Aware Multi-View Contrastive Learning for Patient-Independent ECG Reconstruction
Youssef Youssef, Jitin Singla
Time Series Multimodal
  • Introduction of a novel framework for ECG reconstruction that incorporates pathology-aware embeddings.
  • Achieved a 76% reduction in RMSE compared to state-of-the-art models on the PTB-XL dataset.
  • Demonstrated robust cross-dataset generalization, enhancing the reliability of ECG reconstruction across different populations.
  • Addresses the limitations of standard deep learning methods by focusing on pathological rather than anatomical variations.
Read more
Epistemic Generative Adversarial Networks
Muhammad Mubashar, Fabio Cuzzolin
Generative Models Theory Interpretability
  • Introduction of Epistemic Generative Adversarial Networks (EGANs) to enhance output diversity in GANs.
  • Utilization of Dempster-Shafer theory for a novel GAN loss function that incorporates uncertainty quantification.
  • Architectural modifications to the generator for pixel-wise mass function prediction, improving sample diversity.
  • Experimental results show significant improvements in generation variability and uncertainty modeling.
Read more
Gradient-Informed Temporal Sampling Improves Rollout Accuracy in PDE Surrogate Training
Wenshuo Wang, Fan Zhang
Optimization Time Series Theory
  • Introduces Gradient-Informed Temporal Sampling (GITS) for improved data selection in PDE surrogate training.
  • GITS optimizes local gradient information and temporal coverage to enhance model performance.
  • Demonstrates lower rollout error across multiple PDE systems and neural architectures compared to baseline methods.
  • Ablation studies confirm the necessity of both optimization objectives in GITS.
Read more
BoundAD: Boundary-Aware Negative Generation for Time Series Anomaly Detection
Xiancheng Wang, Lin Wang, Zhibo Zhang, Rui Wang, Minghang Zhao
Time Series Reinforcement Learning Optimization
  • Introduces a reconstruction-driven boundary negative generation framework for TSAD.
  • Utilizes reinforcement learning to adaptively control the generation of hard negatives.
  • Improves anomaly representation learning by focusing on boundary-aware negative samples.
  • Achieves competitive detection performance on benchmark datasets.
Read more
Conflict-Free Policy Languages for Probabilistic ML Predicates: A Framework and Case Study with the Semantic Router DSL
Xunzhuo Liu, Hao Wu, Huamin Chen, Bowei He, Xue Liu
Large Language Models Theory Interpretability
  • Introduces a new framework (ProbPol) for conflict detection in policy languages using probabilistic ML signals.
  • Defines three new conflict types specific to probabilistic predicates and organizes them in a decidability hierarchy.
  • Proposes a method to eliminate co-firing of embedding signals using temperature-scaled softmax.
  • Implements detection and prevention mechanisms in the Semantic Router DSL for practical application.
Read more
Can Blindfolded LLMs Still Trade? An Anonymization-First Framework for Portfolio Optimization
Joohyoung Jeon, Hongchul Lee
Large Language Models Reinforcement Learning Graph Learning
  • Introduces an anonymization protocol to prevent memorization biases in LLM trading agents.
  • Develops a multi-agent system where specialized LLMs assess stocks independently and provide reasoning.
  • Proposes a Semantic Graph Encoder to learn inter-stock relationships under anonymization.
  • Demonstrates rigorous validation of LLM signals to ensure predictive power and mitigate biases.
Read more
Position: Spectral GNNs Are Neither Spectral Nor Superior for Node Classification
Qin Jiang, Chengjia Wang, Michael Lones, Dongdong Chen, Wei Pang
Graph Learning Theory
  • Spectral GNNs do not capture the graph spectrum meaningfully.
  • Commonly used graph Fourier bases are not true Fourier bases.
  • Polynomial approximations in Spectral GNNs are theoretically flawed.
  • The effectiveness of Spectral GNNs is primarily due to message-passing dynamics.
Read more
Path-Constrained Mixture-of-Experts
Zijin Gu, Tatiana Likhomanenko, Vimal Thilak, Jason Ramapuram, Navdeep Jaitly
NLP Large Language Models Efficient ML
  • PathMoE constrains the expert path space by sharing router parameters across consecutive layers.
  • The method shows consistent performance improvements on language modeling tasks compared to independent routing.
  • PathMoE eliminates the need for auxiliary load balancing losses while maintaining balanced expert utilization.
  • Tokens following the same expert path exhibit clustering by linguistic function, enhancing interpretability.
Read more
Detection Is Cheap, Routing Is Learned: Why Refusal-Based Alignment Evaluation Fails
Gregory N. Frank
NLP Large Language Models Theory
  • Detection of sensitive concepts is trivial, but routing through behavioral policies is complex and model-specific.
  • Surgical ablation can effectively remove censorship in many models, leading to accurate outputs.
  • Refusal-based evaluations fail to capture the nuanced steering mechanisms that influence model behavior.
  • Different labs organize political and safety representations in distinct geometries, affecting model outputs.
Read more
InfoMamba: An Attention-Free Hybrid Mamba-Transformer Model
Youjin Wang, Jiaqiao Zhao, Rong Fu, Run Zhou, Ruizhe Zhang, Jiani Liang, Suisuai Cao, Feng Zhou
Efficient ML Computer Vision NLP
  • Introduces InfoMamba, an attention-free hybrid architecture combining SSMs and a global filtering layer.
  • Develops a consistency boundary analysis to identify regimes for effective modeling of local and global interactions.
  • Implements information-maximizing fusion (IMF) to enhance the integration of global context with local details.
  • Achieves strong performance improvements over state-of-the-art Transformer and SSM baselines.
Read more
Transformers Learn Robust In-Context Regression under Distributional Uncertainty
Hoang T. H. Cao, Hai D. V. Trinh, Tho Quan, Lan V. Truong
Theory
  • Transformers can perform in-context learning for linear regression under non-Gaussian and non-i.i.d. conditions.
  • The study systematically evaluates the impact of distributional shifts on Transformer performance.
  • Transformers consistently match or outperform classical regression methods like OLS and Ridge regression.
  • The findings suggest that Transformers exhibit robustness and adaptability beyond traditional estimators.
Read more
DriftGuard: Mitigating Asynchronous Data Drift in Federated Learning
Yizhou Han, Di Wu, Blesson Varghese
Federated Learning
  • DriftGuard effectively mitigates asynchronous data drift in Federated Learning.
  • The framework utilizes a Mixture-of-Experts architecture to separate global and local parameters.
  • It implements a two-level retraining mechanism that balances accuracy and computational overhead.
  • DriftGuard reduces total retraining costs by up to 83% while maintaining high accuracy.
Read more
Federated Distributional Reinforcement Learning with Distributional Critic Regularization
David Millard, Cecilia Alm, Rashid Ali, Pengcheng Shi, Ali Baheri
Reinforcement Learning Federated Learning Robotics
  • Introduction of Federated Distributional Reinforcement Learning (FedDistRL) to address limitations of traditional FRL methods.
  • Development of TR-FedDistRL, which uses a risk-aware Wasserstein barycenter for critic updates.
  • Demonstration of empirical improvements in safety metrics and reduced mean-smearing.
  • Theoretical stability results for the constrained critic update process.
Read more
On Optimizing Multimodal Jailbreaks for Spoken Language Models
Aravind Krishnan, Karolina Stańczak, Dietrich Klakow
Multimodal Audio & Speech Optimization
  • Introduction of JAMA, a joint multimodal optimization framework for jailbreaking SLMs.
  • Demonstrated 1.5x to 10x improvement in jailbreak success rates over unimodal methods.
  • Proposed SAMA as a faster alternative to JAMA with similar effectiveness.
  • Highlighted the need for robust defenses against multimodal adversarial attacks.
Read more
Enhancing Multi-Corpus Training in SSL-Based Anti-Spoofing Models: Domain-Invariant Feature Extraction
Anh-Tuan Dao, Driss Matrouf, Mickael Rouvier, Nicholas Evans
Audio & Speech
  • Multi-corpus training can lead to performance degradation in spoofing detection due to dataset-specific biases.
  • The proposed IDFE framework effectively reduces corpus-specific information in embeddings, enhancing generalization.
  • The framework achieves a 20% reduction in average EER across multiple datasets compared to baseline models.
  • The study highlights the importance of addressing dataset biases for improving the reliability of spoofing detection systems.
Read more
Data-efficient pre-training by scaling synthetic megadocs
Konwoo Kim, Suhas Kotha, Yejin Choi, Tatsunori Hashimoto, Nick Haber, Percy Liang
NLP Large Language Models Efficient ML
  • Synthetic data augmentation can improve pre-training efficiency when real data is scarce.
  • Megadocs, formed by stitching or stretching documents, enhance loss scaling and model performance.
  • Data efficiency improves from 1.48× to 1.80× with optimal synthetic generation strategies.
  • The benefits of synthetic data algorithms compound with existing data-efficient strategies like ensembling.
Read more
Domain-informed explainable boosting machines for trustworthy lateral spread predictions
Cheng-Hsi Hsiao, Krishna Kumar, Ellen M. Rathje
Interpretability
  • Introduces a domain-informed framework to enhance the reliability of EBMs in predicting lateral spreading.
  • Corrects non-physical relationships in EBMs by modifying shape functions based on domain knowledge.
  • Demonstrates the application of the framework on the 2011 Christchurch earthquake dataset.
  • Achieves more physically consistent predictions with an acceptable trade-off in accuracy.
Read more
Translation Invariance of Neural Operators for the FitzHugh-Nagumo Model
Luca Pellegrini
Theory Efficient ML Time Series
  • Introduces a novel training strategy exploiting translation invariance in the FHN model.
  • Benchmarks seven NO architectures, revealing performance trade-offs in training and inference.
  • CNOs perform well on translated dynamics but require higher training costs.
  • FNOs achieve low training error but have high inference time and less accuracy on translated dynamics.
Read more
Learning to Reason with Curriculum I: Provable Benefits of Autocurriculum
Nived Rajaraman, Audrey Huang, Miro Dudik, Robert Schapire, Dylan J. Foster, Akshay Krishnamurthy
NLP Large Language Models Reinforcement Learning Theory Efficient ML
  • Autocurriculum improves training efficiency for reasoning models by adaptively selecting prompts.
  • The proposed method reduces the number of required reasoning demonstrations exponentially compared to non-adaptive approaches.
  • In reinforcement learning, autocurriculum decouples computational costs from model accuracy, enhancing training efficiency.
  • The study provides a theoretical framework for understanding the benefits of autocurriculum in language model training.
Read more
Attention Sinks Induce Gradient Sinks
Yihong Chen, Quanming Yao
Large Language Models Theory Optimization
  • Introduces the concept of gradient sinks as a mechanism linking attention sinks and massive activations.
  • Demonstrates that attention sinks induce pronounced gradient concentration during backpropagation.
  • Shows that massive activations can be interpreted as an adaptive response to gradient pressure.
  • Presents V-scale, a modification that adjusts backpropagated gradients, leading to preserved attention sinks and suppressed massive activations.
Read more
Automatic Configuration of LLM Post-Training Pipelines
Channe Chwa, Xinle Wu, Yao Lu
Large Language Models Reinforcement Learning Optimization
  • Introduction of AutoPipe, a budget-aware framework for LLM post-training configuration selection.
  • Development of a dataset-conditioned ranking surrogate that enhances guidance across datasets.
  • Implementation of an online adaptation mechanism using Gaussian-process modeling and an early-stop predictor.
  • Demonstrated effectiveness of AutoPipe in reducing computational costs while achieving competitive performance.
Read more
TimeAPN: Adaptive Amplitude-Phase Non-Stationarity Normalization for Time Series Forecasting
Yue Hu, Jialiang Tang, Siwei Yu, Baosheng Yu, Jing Zhang, Dacheng Tao
Time Series
  • TimeAPN effectively models non-stationary factors in time and frequency domains.
  • The framework captures rapid changes in amplitude and phase for improved forecasting.
  • TimeAPN integrates amplitude and phase information through an adaptive normalization mechanism.
  • Extensive experiments show significant improvements in forecasting accuracy over state-of-the-art methods.
Read more
Fundamental Limits of Neural Network Sparsification: Evidence from Catastrophic Interpretability Collapse
Dip Roy, Rajiv Misra, Sanjay Kumar Singh
Interpretability
  • Extreme sparsification leads to significant interpretability collapse, despite stable global representation quality.
  • Local interpretability metrics deteriorate while global disentanglement metrics remain stable under compression.
  • The phenomenon of interpretability collapse is intrinsic to the sparsification process, not dependent on specific algorithms or training durations.
  • Dead neuron rates increase with dataset complexity, indicating greater challenges for real-world applications.
Read more
CausalRM: Causal-Theoretic Reward Modeling for RLHF from Observational User Feedbacks
Hao Wang, Licheng Pan, Zhichao Chen, Chunyuan Zheng, Zhixuan Chu, Xiaoxi Li, Yuan Lu, Xinggao Liu, Haoxuan Li, Zhouchen Lin
Reinforcement Learning Large Language Models Theory
  • CausalRM provides a scalable alternative to traditional reward modeling by utilizing observational user feedback.
  • The framework addresses noise and bias in observational feedback through innovative loss modeling and reweighting techniques.
  • Extensive experiments confirm CausalRM's effectiveness across multiple LLMs and datasets, leading to substantial performance gains.
  • The work formalizes the problem of observational reward modeling, highlighting its significance in RLHF applications.
Read more
Integrating Explainable Machine Learning and Mixed-Integer Optimization for Personalized Sleep Quality Intervention
Mahfuz Ahmed Anik, Mohsin Mahmud Topu, Azmine Toushik Wasi, Md Isfar Khan, MD Manjurul Ahsan
Optimization Interpretability
  • Proposes a novel framework integrating explainable machine learning and mixed-integer optimization for sleep interventions.
  • Achieves high predictive accuracy and identifies key behavioral factors influencing sleep quality.
  • Generates personalized recommendations that consider individual constraints and the feasibility of behavioral changes.
  • Demonstrates a trade-off between expected improvement and intervention intensity, highlighting diminishing returns.
Read more
A Model Ensemble-Based Post-Processing Framework for Fairness-Aware Prediction
Zhouting Zhao, Tin Lok James Ng
Theory Optimization Efficient ML
  • Proposes a model ensemble-based framework for fairness-aware predictions.
  • Framework is model-agnostic and applicable across various learning tasks.
  • Enhances fairness while minimally affecting predictive accuracy.
  • Extends applicability to survival analysis, a critical area in machine learning.
Read more
SOL-ExecBench: Speed-of-Light Benchmarking for Real-World GPU Kernels Against Hardware Limits
Edward Lin, Sahil Modi, Siva Kumar Sastry Hari, Qijing Huang, Zhifan Ye, Nestor Qin, Fengzhe Zhou, Yuan Zhang, Jingquan Wang, Sana Damani, Dheeraj Peri, Ouye Xie, Aditya Kane, Moshe Maor, Michael Behar, Triston Cao, Rishabh Mehta, Vartika Singh, Vikram Sharma Mailthody, Terry Chen, Zihao Ye, Hanfeng Chen, Tianqi Chen, Vinod Grover, Wei Chen, Wei Liu, Eric Chung, Luis Ceze, Roger Bringmann, Cyril Zeller, Michael Lightstone, Christos Kozyrakis, Humphrey Shi
Optimization Efficient ML Generative Models
  • SOL-ExecBench benchmarks GPU kernels against hardware Speed-of-Light limits rather than software baselines.
  • The benchmark includes 235 CUDA kernel optimization problems from diverse AI applications.
  • SOL Score measures the performance gap closure between a kernel and analytically derived SOL bounds.
  • A sandboxed evaluation harness ensures reliable and reproducible benchmarking results.
Read more
Baguan-TS: A Sequence-Native In-Context Learning Model for Time Series Forecasting with Covariates
Linxiao Yang, Xue Jiang, Gezheng Xu, Tian Zhou, Min Yang, ZhaoYang Zhu, Linyuan Geng, Zhipeng Zeng, Qiming Chen, Xinyue Gu, Rong Jin, Liang Sun
Time Series
  • Baguan-TS enables in-context learning directly on raw multivariate time series without feature engineering.
  • The model employs a 3D Transformer architecture that attends across temporal, variable, and context axes.
  • A Y-space retrieval-based calibration module enhances model stability and forecasting accuracy.
  • The context-overfitting strategy improves robustness by balancing denoising and selection of relevant support examples.
Read more
One-Step Sampler for Boltzmann Distributions via Drifting
Wenhan Cao, Keyu Yan, Lin Zhao
Generative Models Optimization Efficient ML
  • Introduces a drifting-based framework for sampling Boltzmann distributions.
  • Develops a one-step neural generator that simplifies the sampling process.
  • Utilizes Gaussian-smoothed score fields to project samples towards target distributions.
  • Achieves low error rates on complex target distributions, demonstrating efficiency.
Read more
Escaping Offline Pessimism: Vector-Field Reward Shaping for Safe Frontier Exploration
Amirhossein Roknilamouki, Arnob Ghosh, Eylem Ekici, Ness B. Shroff
Reinforcement Learning Robotics Theory
  • Introduces vector-field reward shaping to enhance exploration in offline RL.
  • Combines gradient-alignment and rotational-flow terms for effective boundary exploration.
  • Theoretical analysis supports the efficacy of the proposed reward structure.
  • Empirical results demonstrate successful navigation of uncertainty boundaries.
Read more
Evaluating Model-Free Policy Optimization in Masked-Action Environments via an Exact Blackjack Oracle
Kevin Song
Reinforcement Learning Optimization Theory
  • Development of an exact dynamic programming oracle for blackjack, providing a rigorous benchmark for policy optimization.
  • Evaluation of three model-free optimizers (REINFORCE, SPSA, CEM) in recovering optimal policies under dynamically masked actions.
  • REINFORCE demonstrated the highest sample efficiency, but all methods faced significant policy-level errors.
  • Establishment of a minimum-bet optimality theorem, confirming that optimal betting strategies under no-count conditions lead to minimum wagers.
Read more
Benchmarking Reinforcement Learning via Stochastic Converse Optimality: Generating Systems with Known Optimal Policies
Sinan Ibrahim, Grégoire Ouerdane, Hadi Salloum, Henni Ouerdane, Stefan Streif, Pavel Osinenko
Reinforcement Learning Theory Optimization
  • Introduction of a benchmarking framework based on stochastic converse optimality.
  • Systematic generation of benchmark families with known optimal policies.
  • Validation through the automatic construction of diverse environments.
  • Provision of a reproducible foundation for evaluating RL algorithms against certified optima.
Read more
Lightweight Adaptation for LLM-based Technical Service Agent: Latent Logic Augmentation and Robust Noise Reduction
Yi Yu, Junzhuo Ma, Chenghuang Shen, Xingyan Liu, Jing Gu, Hangyi Sun, Guangquan Hu, Jianfeng Liu, Weiting Liu, Mingyue Pu, Yu Wang, Zhengdong Xiao, Rui Xie, Longjiu Luo, Qianrong Wang, Gurong Cui, Honglin Qiao, Wenlian Lu
NLP Large Language Models Reinforcement Learning
  • Introduces Latent Logic Augmentation to enhance decision-making in LLMs.
  • Develops a Multiple Ground Truths dataset to reduce training noise and capture semantic diversity.
  • Presents a Hybrid Reward mechanism for efficient reinforcement learning.
  • Demonstrates improved stability and performance in real-world Cloud service tasks.
Read more
From ex(p) to poly: Gaussian Splatting with Polynomial Kernels
Joerg H. Mueller, Martin Winter, Markus Steinberger
Computer Vision Efficient ML
  • Introduction of an N-th-order polynomial kernel approximation for Gaussian Splatting.
  • Significant reduction in computational overhead while maintaining compatibility with existing datasets.
  • Derivation of tighter bounding radii for aggressive splat culling, enhancing performance.
  • Formal proof of invariance of anti-aliasing normalization factors for arbitrary kernel functions.
Read more
AIMER: Calibration-Free Task-Agnostic MoE Pruning
Zongfang Liu, Shengkun Tang, Yifan Shen, Huan Wang, Xin Yuan
NLP Large Language Models Efficient ML
  • AIMER addresses the calibration dependence issue in task-agnostic MoE expert pruning.
  • The proposed method allows for expert ranking without relying on calibration data.
  • AIMER achieves superior expert stratification compared to traditional weight magnitude methods.
  • The method demonstrates competitive performance across multiple benchmarks and models.
Read more
The Causal Uncertainty Principle: Manifold Tearing and the Topological Limits of Counterfactual Interventions
Rui Wu, Hong Xie, Yongjun Li
Theory Generative Models Optimization
  • Establishes the topological limits of counterfactual interventions in continuous spaces.
  • Introduces the Counterfactual Event Horizon, defining critical transport distances for causal interventions.
  • Demonstrates that extreme interventions can lead to finite-time singularities (Manifold Tearing).
  • Proposes Geometry-Aware Causal Flow (GACF) to manage geometric entropy during counterfactual generation.
Read more
SLEA-RL: Step-Level Experience Augmented Reinforcement Learning for Multi-Turn Agentic Training
Prince Zizhuang Wang, Shuli Jiang
Reinforcement Learning Large Language Models Robotics
  • SLEA-RL retrieves experiences at each decision step, enhancing adaptability in multi-turn tasks.
  • The framework employs observation clustering for efficient experience retrieval and generalization.
  • A self-evolving experience library maintains quality through score-based admission and rate-limited extraction.
  • SLEA-RL shows superior performance on benchmarks like ALFWorld and WebShop compared to traditional RL methods.
Read more
A foundation model for electrodermal activity data
Leonardo Alchieri, Matteo Garzon, Lidia Alecci, Francesco Bombassei De Bona, Martin Gjoreski, Giovanni De Felice, Silvia Santini
Time Series
  • Introduction of EDAMAME, a large-scale EDA dataset from 24 public sources.
  • Development of UME, the first foundation model specifically for EDA data.
  • UME outperforms baseline models in 80% of scenarios while being 20 times more computationally efficient.
  • Challenges in EDA modeling are acknowledged, indicating the need for further research.
Read more
SCALE: Scalable Conditional Atlas-Level Endpoint transport for virtual cell perturbation prediction
Shuizhou Chen, Lang Yu, Kedu Jin, Songming Zhang, Hao Wu, Wenxuan Huang, Sheng Xu, Quan Qian, Qin Chen, Lei Bai, Siqi Sun, Zhangyang Gao
Generative Models Efficient ML Theory
  • SCALE significantly improves the efficiency of virtual cell perturbation prediction through a specialized training and inference framework.
  • The model utilizes a set-aware flow architecture for stable and biologically faithful predictions of perturbation effects.
  • SCALE achieves notable performance improvements over existing models on the Tahoe-100M benchmark.
  • The research emphasizes the need for co-designing scalable systems and biologically grounded evaluation metrics in virtual cell modeling.
Read more
AcceRL: A Distributed Asynchronous Reinforcement Learning and World Model Framework for Vision-Language-Action Models
Chengxuan Lu, Shukuan Wang, Yanjie Li, Wei Liu, Shiji Jin, Fuyuan Qian, Peiming Li, Baigui Sun, Yang Liu
Reinforcement Learning Multimodal Robotics
  • AcceRL eliminates synchronization barriers by decoupling training, inference, and rollouts.
  • The framework integrates a trainable world model, enhancing sample efficiency by 200x.
  • AcceRL achieves state-of-the-art performance on the LIBERO benchmark.
  • The architecture demonstrates super-linear scaling in throughput and efficient hardware utilization.
Read more
Adaptive Regime-Aware Stock Price Prediction Using Autoencoder-Gated Dual Node Transformers with Reinforcement Learning Control
Mohammad Al Ridhawi, Mahtab Haj Ali, Hussein Al Osman
Reinforcement Learning Time Series
  • Introduces an adaptive framework for stock price prediction that identifies market regime shifts.
  • Utilizes an autoencoder for weakly supervised anomaly detection based on reconstruction errors.
  • Employs dual node transformers for specialized processing of stable and volatile market conditions.
  • Incorporates a Soft Actor-Critic reinforcement learning controller for dynamic threshold adjustment.
Read more
Context Bootstrapped Reinforcement Learning
Saaket Agashe, Jayanth Srinivasa, Gaowen Liu, Ramana Kompella, Xin Eric Wang
Reinforcement Learning Large Language Models
  • CBRL effectively addresses exploration inefficiency in RLVR by using in-context learning.
  • The method employs a curriculum-based approach for injecting few-shot demonstrations.
  • CBRL shows consistent performance improvements across various tasks and model families.
  • The approach is algorithm-agnostic, yielding benefits regardless of the underlying reinforcement learning algorithm.
Read more
Attack by Unlearning: Unlearning-Induced Adversarial Attacks on Graph Neural Networks
Jiahao Zhang, Yilong Wang, Suhang Wang
Graph Learning Optimization Theory
  • Introduction of unlearning corruption attacks that exploit the unlearning process in GNNs.
  • Formulation of the attack as a bi-level optimization problem to address black-box challenges.
  • Demonstration of significant accuracy degradation in GNNs due to carefully crafted unlearning requests.
  • Highlighting the stealthy nature of the attack, which remains undetected during training.
Read more
Hierarchical Latent Structure Learning through Online Inference
Ines Aitsahalia, Kiyohito Iigaya
Theory Efficient ML Time Series
  • HOLMES integrates hierarchical representation with online inference for latent structure learning.
  • The model utilizes a nested Chinese Restaurant Process prior and sequential Monte Carlo methods.
  • HOLMES achieves compact representations that support one-shot transfer to higher-level categories.
  • The model outperforms flat models in context-dependent tasks with nested temporal structures.
Read more
Binary Latent Protein Fitness Landscapes for Quantum Annealing Optimization
Truong-Son Hy
Optimization
  • Introduction of Q-BioLat framework for protein fitness landscape optimization.
  • Utilizes pretrained protein language models to create binary latent representations.
  • Models protein fitness as a QUBO problem for efficient combinatorial optimization.
  • Demonstrates effective identification of high-fitness protein variants on the ProteinGym benchmark.
Read more
Taming Epilepsy: Mean Field Control of Whole-Brain Dynamics
Ming Li, Ting Gao, Jingqiao Duan
Graph Learning Optimization Theory
  • Introduction of the GK-MFG framework for controlling epileptic seizures.
  • Integration of Reservoir Computing with graph-theoretic modeling for effective neural dynamics control.
  • Use of graph Laplacian constraints to respect the brain's functional topology.
  • Demonstration of robust seizure suppression in high-dimensional networks.
Read more
Variational Phasor Circuits for Phase-Native Brain-Computer Interface Classification
Dibakar Sigdel
Theory Efficient ML Time Series
  • Introduction of the Variational Phasor Circuit (VPC) as a phase-native learning architecture.
  • VPC replaces dense weight matrices with trainable phase shifts, enhancing parameter efficiency.
  • Demonstrated competitive performance on synthetic BCI benchmarks with fewer parameters than traditional methods.
  • VPC serves as a bridge between classical oscillatory signal processing and future quantum systems.
Read more
A Family of Adaptive Activation Functions for Mitigating Failure Modes in Physics-Informed Neural Networks
Krishna Murari
Theory Optimization Efficient ML
  • Introduction of adaptive wavelet-based activation functions to enhance PINNs.
  • Improved training stability and expressive power compared to traditional activation functions.
  • Evaluation across multiple classes of PDEs demonstrating superior performance.
  • Validation against various models, including baseline PINNs and transformer architectures.
Read more
Self-Tuning Sparse Attention: Multi-Fidelity Hyperparameter Optimization for Transformer Acceleration
Arundhathi Dev, Justin Zhan
NLP Large Language Models Efficient ML
  • AFBS-BO automates the hyperparameter tuning process for sparse attention, eliminating the need for manual grid search.
  • The method achieves a 3.4x speedup in hyperparameter discovery and requires 8.8x fewer evaluations than traditional methods.
  • Configurations discovered by AFBS-BO outperform existing sparse attention baselines while closely matching the quality of dense attention.
  • The framework leverages multi-fidelity evaluations to efficiently explore the hyperparameter space.
Read more
Discovering Decoupled Functional Modules in Large Language Models
Yanke Yu, Jin Li, Ying Sun, Ping Li, Zhefeng Wang, Yi Zheng
Large Language Models Interpretability
  • Introduces the function module discovery problem in LLMs, addressing a critical gap in interpretability research.
  • Develops the ULCMOD framework with a novel objective function and IterD algorithm for effective module identification.
  • Demonstrates superior performance in module discovery compared to existing clustering methods.
  • Provides insights into the organization of functions within LLMs, highlighting comprehensiveness and hierarchy.
Read more
HEP Statistical Inference for UAV Fault Detection: CLs, LRT, and SBI Applied to Blade Damage
Khushiyant
Robotics Time Series Theory
  • Introduction of HEP statistical methods to UAV fault detection.
  • Unified inference framework providing binary detection, false alarm control, and fault characterization.
  • Significant performance improvement over traditional methods in detecting blade damage.
  • SNPE offers calibrated uncertainty estimates for fault severity.
Read more
Physics-Aware Machine Learning for Seismic and Volcanic Signal Interpretation
William Thorossian
Time Series
  • Machine learning is becoming essential for processing seismic and volcanic data, but must adapt to domain shifts.
  • Models need to provide uncertainty estimates to support decision-making in monitoring.
  • Integrating classical signal processing with ML can enhance performance and reliability.
  • Evaluation protocols should reflect the challenges of transferring models across different regions and conditions.
Read more
Large-Scale 3D Ground-Motion Synthesis with Physics-Inspired Latent Operator Flow Matching
Yaozhong Shi, Grigorios Lavrentiadis, Konstantinos Tsalouchidis, Zachary E. Ross, David McCallen, Caifeng Zou, Kamyar Azizzadenesheli, Domniki Asimaki
Generative Models Time Series Efficient ML
  • GMFLOW achieves a 10,000-fold speedup in generating ground-motion time histories compared to traditional methods.
  • The framework operates in two stages: low-frequency wavefield generation followed by high-resolution reconstruction.
  • GMFLOW supports zero-shot super-resolution, allowing for flexible spatial resolution in predictions.
  • The method is validated on simulated earthquake scenarios in the San Francisco Bay Area.
Read more
Online Learning and Equilibrium Computation with Ranking Feedback
Mingyang Liu, Yongshan Chen, Zhiyuan Fan, Gabriele Farina, Asuman Ozdaglar, Kaiqing Zhang
Theory Optimization
  • Introduces a ranking-based online learning model that does not rely on numeric utility feedback.
  • Establishes that sublinear regret is unattainable with instantaneous utility rankings.
  • Develops algorithms achieving sublinear regret under specific conditions related to utility variation.
  • Demonstrates that the algorithms lead to approximate coarse correlated equilibria in game-theoretic settings.
Read more
AGRI-Fidelity: Evaluating the Reliability of Listenable Explanations for Poultry Disease Detection
Sindhuja Madabushi, Arda Dogan, Jonathan Liu, Dian Chen, Dong S. Ha, Sook Shin, Sam H. Noh, Jin-Hee Cho
Audio & Speech Interpretability
  • AGRI-Fidelity introduces a reliability-centered evaluation paradigm for XAI in bioacoustic disease detection.
  • The framework integrates cross-model consensus with fidelity-based causal validation to quantify explanation stability.
  • A novel permutation-based null construction is developed to estimate empirical False Discovery Rates, suppressing irrelevant artifacts.
  • Extensive experiments show that AGRI-Fidelity consistently differentiates between meaningful signals and domain-irrelevant artifacts.
Read more
From Language to Action in Arabic: Reliable Structured Tool Calling via Data-Centric Fine-Tuning
Omer Nacar, Deema Alquffari, Saleh Alsharideh, Adeem AlOtaibi, Abdulaziz Alabdulkarim, Leen Alhazmi, Nada Alomar, Wareef Alzubaidi, Nada Alsultan, Ahmed Alrabghi, Demah Alhoshan, Rana Alsayyari, Hamed Alruwaili, Albaraa Jaafar, Khaled Alusmani, Abdulaziz Alsohimy, Munirah Alsubaie, Shahd Aldukhayil, Arwa Alali, Yazeed BinShihah, Razan Alsulaymi, Nourah Alhumaid, Razan Abdulsalam, Reem Alamoudi, Mohammed Alkhalifa
NLP Large Language Models
  • Introduction of AISA-AR-FunctionCall framework for Arabic function calling.
  • Significant reduction in parse failures and improvement in function name accuracy.
  • Error analysis reveals distinct challenges in serialization stability and reasoning.
  • Exploration of reasoning-augmented LoRA variant for enhanced decision-making.
Read more
Towards Infinitely Long Neural Simulations: Self-Refining Neural Surrogate Models for Dynamical Systems
Qi Liu, Laure Zanna, Joan Bruna
Generative Models Theory Time Series
  • Introduces a mathematical framework for balancing short-term accuracy and long-term consistency in neural simulations.
  • Proposes the Self-refining Neural Surrogate model (SNS) as a hyperparameter-free solution.
  • Demonstrates the effectiveness of SNS in high-fidelity simulations of dynamical systems over long time horizons.
  • Addresses the issue of distribution drift in autoregressive models, enhancing their reliability.
Read more
Approximate Subgraph Matching with Neural Graph Representations and Reinforcement Learning
Kaiyang Li, Shihao Ji, Zhipeng Cai, Wei Li
Reinforcement Learning Graph Learning Optimization
  • RL-ASM is the first approach to apply reinforcement learning to approximate subgraph matching.
  • The method utilizes a Graph Transformer to fully exploit graph information for improved matching.
  • RL-ASM optimizes node pair selection based on long-term rewards rather than greedy heuristics.
  • Extensive experiments show RL-ASM outperforms traditional ASM techniques in various scenarios.
Read more
Enactor: From Traffic Simulators to Surrogate World Models
Yash Ranjan, Rahul Sengupta, Anand Rangarajan, Sanjay Ranka
Generative Models Reinforcement Learning Robotics
  • Introduces a transformer-based generative model for traffic simulation.
  • Addresses limitations of traditional microsimulators in capturing realistic actor interactions.
  • Demonstrates improved performance in generating long-horizon, physically consistent trajectories.
  • Requires fewer training samples compared to traditional agent-centric approaches.
Read more
From Inference Efficiency to Embodied Efficiency: Revisiting Efficiency Metrics for Vision-Language-Action Models
Zhuofan Li, Hongkun Yang, Zhenyang Chen, Yangxuan Chen, Yingyan (Celine) Lin, Chaojian Li
Multimodal Robotics Efficient ML
  • Conventional efficiency metrics do not capture the embodied performance of VLA models.
  • Embodied efficiency metrics reveal hidden performance differences in learned action policies.
  • Common adaptation methods yield limited improvements and may introduce trade-offs.
  • Reducing computational costs does not guarantee improved embodied execution efficiency.
Read more
FlowMS: Flow Matching for De Novo Structure Elucidation from Mass Spectra
Jianan Nie, Peng Gao
Generative Models Graph Learning
  • FlowMS is the first discrete flow matching framework for de novo molecular generation from mass spectra.
  • It achieves state-of-the-art performance on 5 out of 6 metrics on the NPLIB1 benchmark.
  • FlowMS demonstrates a 9.15% top-1 accuracy, surpassing previous models like DiffMS and MS-BART.
  • The framework effectively enforces chemical formula constraints during molecular generation.
Read more
Procedural Generation of Algorithm Discovery Tasks in Machine Learning
Alexander D. Goldie, Zilin Wang, Adrian Hayler, Deepak Nathani, Edan Toledo, Ken Thampiratwong, Aleksandra Kalisz, Michael Beukman, Alistair Letcher, Shashank Reddy, Clarisse Wibault, Theo Wolf, Charles O'Neill, Uljad Berdica, Nicholas Roberts, Saeed Rahmani, Hannah Erlebach, Roberta Raileanu, Shimon Whiteson, Jakob N. Foerster
Optimization Reinforcement Learning Theory
  • DiscoGen generates over 400 million diverse algorithm discovery tasks, addressing limitations of existing benchmarks.
  • The framework ensures principled evaluation by separating meta-train and meta-test datasets.
  • DiscoBench provides a curated set of tasks for evaluating algorithm discovery agents (ADAs).
  • The methodology supports an ADA optimization loop, enhancing the iterative development of algorithms.
Read more
Contextual Preference Distribution Learning
Benjamin Hudson, Laurent Charlin, Emma Frejinger
Optimization
  • Introduces a method for learning context-dependent preference distributions in decision-making problems.
  • Utilizes a sequential learning-and-optimization pipeline to address human preference uncertainty.
  • Achieves maximum likelihood estimates with desirable statistical properties.
  • Demonstrates substantial reductions in post-decision surprise in a ridesharing context.
Read more