AI-generated summaries

Today's ML research,
without the noise.

Daily summaries of the latest machine learning papers from arXiv, processed every 8 hours.

65 Papers today
8h Update frequency
7 Days of history
Step-level Denoising-time Diffusion Alignment with Multiple Objectives
Qi Zhang, Dawei Wang, Shaofeng Zou
Reinforcement Learning Generative Models Optimization
  • Introduces a retraining-free framework for aligning diffusion models with multiple objectives.
  • Develops a step-level RL formulation that avoids the need for reward gradients and approximation errors.
  • Derives a closed-form solution for the optimal reverse denoising distribution based on single-objective models.
  • Demonstrates superior performance compared to existing denoising-time approaches through extensive experiments.
Read more
FedIDM: Achieving Fast and Stable Convergence in Byzantine Federated Learning through Iterative Distribution Matching
He Yang, Dongyi Lv, Wei Xi, Song Ma, Hanlin Gu, Jizhong Zhao
Federated Learning
  • FedIDM leverages iterative distribution matching for robust and efficient convergence in Byzantine FL.
  • The framework minimizes the impact on model utility even with a high proportion of colluded malicious clients.
  • Empirical evaluations show substantial improvements over existing Byzantine-robust methods.
  • The attack-tolerant condensed data generation effectively counters label-flipping attacks.
Read more
Evaluating Supervised Machine Learning Models: Principles, Pitfalls, and Metric Selection
Xuanyan Liu, Ignacio Cabrera Martin, Marcello Trovati, Xiaolong Xu, Nikolaos Polatidis
Theory
  • Model evaluation is often reduced to a few aggregate metrics, risking misleading conclusions.
  • Common pitfalls in evaluation include data leakage, class imbalance, and inappropriate metric selection.
  • Evaluation should be treated as a decision-oriented and context-dependent process.
  • The paper emphasizes the importance of aligning evaluation methods with operational objectives.
Read more
Gating Enables Curvature: A Geometric Expressivity Gap in Attention
Satwik Bathula, Anand A. Joshi
NLP Large Language Models Theory
  • Gated attention mechanisms enable non-flat geometries, enhancing representational expressivity.
  • Ungated attention is limited to flat statistical manifolds due to its affine structure.
  • Multiplicative gating introduces nonlinear modulation, allowing for richer representation structures.
  • Empirical results indicate gated models perform better on tasks with nonlinear decision boundaries.
Read more
The Devil Is in Gradient Entanglement: Energy-Aware Gradient Coordinator for Robust Generalized Category Discovery
Haiyang Zheng, Nan Pu, Yaqi Cai, Teng Long, Wenjing Li, Nicu Sebe, Zhun Zhong
Computer Vision Optimization Theory
  • Identifies Gradient Entanglement as a critical issue limiting GCD performance.
  • Proposes EAGC, a plug-and-play module that effectively mitigates GE.
  • Includes AGA and EEP components to enhance gradient optimization.
  • Achieves new state-of-the-art results in GCD across various benchmarks.
Read more
Graph-Based Fraud Detection with Dual-Path Graph Filtering
Wei He, Wensheng Gan, Philip S. Yu
Graph Learning
  • DPF-GFD introduces a dual-path filtering approach to enhance fraud detection in graph data.
  • The model effectively addresses challenges such as relation camouflage and high heterophily.
  • It employs a beta wavelet-based operator for structural pattern extraction and a similarity graph for feature representation.
  • The method shows improved performance on real-world financial fraud detection datasets.
Read more
Portfolio Optimization Proxies under Label Scarcity and Regime Shifts via Bayesian and Deterministic Students under Semi-Supervised Sandwich Training
Adhiraj Chattopadhyay
Optimization
  • Introduces a teacher-student learning framework for portfolio optimization using CVaR as a supervisory signal.
  • Utilizes Bayesian Neural Networks (BNNs) to provide uncertainty-aware predictions and mitigate overfitting in low-data settings.
  • Demonstrates implicit reduction in trading turnover, achieving a 50% decrease compared to deterministic models without explicit constraints.
  • Shows that the learned policies generalize effectively across different market conditions and asset universes.
Read more
Asynchronous Probability Ensembling for Federated Disaster Detection
Emanuel Teixeira Martins, Rodrigo Moreira, Larissa Ferreira Rodrigues Moreira, Rodolfo S. Villaça, Augusto Neto, Flávio de Oliveira Silva
Federated Learning Computer Vision Efficient ML
  • Introduces an asynchronous probability-level aggregation framework for disaster detection.
  • Reduces communication overhead by exchanging class-probability vectors instead of model weights.
  • Enhances collaboration among heterogeneous CNN architectures without requiring synchronization.
  • Integrates ensemble strategies and a knowledge distillation feedback loop for improved accuracy.
Read more
MOONSHOT : A Framework for Multi-Objective Pruning of Vision and Large Language Models
Gabriel Afriat, Xiang Meng, Shibal Ibrahim, Hussein Hazimeh, Rahul Mazumder
Computer Vision Large Language Models Efficient ML
  • MOONSHOT enhances one-shot pruning by optimizing multiple objectives simultaneously.
  • The framework is scalable and efficient, suitable for billion-parameter models.
  • Experimental results show significant improvements in performance and accuracy across various models.
  • The study reveals that different pruning criteria can yield complementary insights into parameter importance.
Read more
Mean Flow Policy Optimization
Xiaoyi Dong, Xi Sheryl Zhang, Jian Cheng
Reinforcement Learning Generative Models Optimization
  • MFPO leverages MeanFlow models to enhance efficiency in online RL.
  • The approach promotes exploration through maximum entropy RL and soft policy iteration.
  • MFPO addresses challenges in action likelihood evaluation and soft policy improvement.
  • Experimental results demonstrate superior performance with reduced computational overhead.
Read more
Does RL Expand the Capability Boundary of LLM Agents? A PASS@(k,T) Analysis
Zhiyuan Zhai, Wenjing Yan, Xiaodan Shao, Xin Wang
Large Language Models Reinforcement Learning Theory
  • Introduces PASS@(k, T), a two-dimensional evaluation metric for LLM agents.
  • Demonstrates that RL expands the capability boundary of LLM agents in compositional tool-use tasks.
  • Finds that supervised fine-tuning can regress capabilities on similar tasks, isolating exploration as a key factor.
  • Mechanistic analysis reveals RL improves strategy selection and information integration.
Read more
From Risk to Rescue: An Agentic Survival Analysis Framework for Liquidation Prevention
Fernando Spadea, Oshani Seneviratne
Optimization Time Series Theory
  • The framework transitions from passive risk prediction to proactive intervention in liquidation prevention.
  • A novel return period metric is introduced to normalize risk across transaction types.
  • The counterfactual optimization loop simulates user actions to minimize required capital for risk mitigation.
  • The system effectively differentiates between significant financial risks and minor 'dust' events.
Read more
Auxiliary Finite-Difference Residual-Gradient Regularization for PINNs
Stavros Kassinos
Theory Optimization
  • Introduces an auxiliary finite-difference regularizer for PINNs that maintains the governing PDE residual in AD form.
  • Demonstrates a trade-off between accuracy of the field and cleanliness of the residual in a controlled Poisson problem study.
  • Implements a body-fitted shell regularizer in a 3D heat-conduction benchmark, improving application-specific quantities.
  • Identifies optimal configurations for regularization and learning rates that enhance model reliability.
Read more
MambaSL: Exploring Single-Layer Mamba for Time Series Classification
Yoo-Min Jung, Leekyung Kim
Time Series
  • MambaSL achieves state-of-the-art performance in time series classification.
  • The framework is guided by four TSC-specific hypotheses that refine Mamba's architecture.
  • A unified benchmarking protocol is established, addressing issues of coverage, fairness, and reproducibility.
  • MambaSL outperforms the second-best method by 1.41% in accuracy across UEA datasets.
Read more
GUI-Perturbed: Domain Randomization Reveals Systematic Brittleness in GUI Grounding Models
Yangyue Wang, Harshvardhan Sikka, Yash Mathur, Tony Zhou, Jinu Nyachhyon, Pranav Guruprasad
Computer Vision NLP Multimodal
  • GUI grounding models show a significant accuracy drop (27-56 pp) when tasked with spatial reasoning.
  • A 70% browser zoom leads to a notable performance degradation across all tested models.
  • Standard training methods do not enhance model robustness and may worsen spatial reasoning abilities.
  • The GUI-Perturbed framework allows for controlled evaluation of grounding robustness by varying visual and instructional conditions.
Read more
Soft $Q(λ)$: A multi-step off-policy method for entropy regularised reinforcement learning using eligibility traces
Pranav Mahajan, Ben Seymour
Reinforcement Learning
  • Introduces Soft Q(λ), a multi-step off-policy method for entropy-regularized reinforcement learning.
  • Develops a novel Soft Tree Backup operator to handle entropy terms across multiple time steps.
  • Eliminates the on-policy bias inherent in traditional n-step soft Q-learning methods.
  • Demonstrates the ability to learn entropy-regularized value functions under arbitrary behavior policies.
Read more
When Less Latent Leads to Better Relay: Information-Preserving Compression for Latent Multi-Agent LLM Collaboration
Yiping Li, Zhiyu An, Wan Du
Large Language Models NLP Efficient ML
  • Introduces Orthogonal Backfill (OBF) to enhance KV compression in multi-agent LLM communication.
  • Achieves a significant reduction in communication costs (79.8%–89.4%) while maintaining competitive performance.
  • Demonstrates that preserving useful information is more critical than merely relaying large amounts of data.
  • Evaluates the method across nine diverse benchmarks, showing superior results in several cases.
Read more
Curvature-Aligned Probing for Local Loss-Landscape Stabilization
Nikita Kiselev, Andrey Grabovoy
Theory Optimization Efficient ML
  • Introduces a unified family of local stabilization criteria for loss-landscape analysis.
  • Proposes a curvature-aligned criterion ∆(D)² that focuses on the top-D eigenspace of the Hessian.
  • Demonstrates that the new criterion preserves the mean-squared rate of traditional methods while improving efficiency.
  • Develops scalable estimators that significantly reduce computational costs compared to direct Monte Carlo methods.
Read more
RPS: Information Elicitation with Reinforcement Prompt Selection
Tao Wang, Jingyao Lu, Xibo Wang, Haonan Huang, Su Yao, Zhiqiang Hu, Xingyan Chen, Enmao Diao
NLP Large Language Models Reinforcement Learning
  • Proposes Reinforcement Prompt Selection (RPS) for adaptive information elicitation in dialogues.
  • Introduces IELegal, a benchmark dataset for evaluating information elicitation in legal contexts.
  • RPS outperforms static prompt baselines, demonstrating the effectiveness of adaptive strategies.
  • Addresses the limitations of existing prompt engineering methods by reducing reliance on static prompts.
Read more
Metric-Aware Principal Component Analysis (MAPCA): A Unified Framework for Scale-Invariant Representation Learning
Michael Leznik
Theory
  • MAPCA provides a unified framework for scale-invariant representation learning by utilizing a metric matrix.
  • The β-family of metrics allows for interpolation between standard PCA and output whitening, addressing the trade-off between scale invariance and variance preservation.
  • Invariant PCA (IPCA) is identified as a special case within the MAPCA framework, showcasing strict scale invariance.
  • Connections to self-supervised learning methods are established, clarifying their underlying metric choices.
Read more
TOPCELL: Topology Optimization of Standard Cell via LLMs
Zhan Song, Yu-Tung Liu, Chen Chen, Guoheng Sun, Jiaqi Yin, Chia-tung Ho, Ang Li, Haoxing Ren, Cunxi Yu
Large Language Models Optimization Generative Models
  • Introduction of TOPCELL, a framework leveraging LLMs for topology optimization in standard cell design.
  • Utilization of Group Relative Policy Optimization (GRPO) for efficient and autonomous topology discovery.
  • Demonstrated superior performance and zero-shot generalization in generating high-quality topologies.
  • Achieved an average speedup of 85.91x compared to traditional exhaustive search methods.
Read more
Quantum-inspired tensor networks in machine learning models
Guillermo Valverde, Igor García-Olaizola, Giannicola Scarpa, Alejandro Pozas-Kerstjens
Theory Efficient ML Interpretability
  • Tensor networks offer a compressed representation of complex data dependencies, improving computational efficiency.
  • They can enhance explainability and privacy in machine learning models compared to traditional neural networks.
  • Two main approaches are discussed: using TNs as learning architectures and as compression strategies for existing models.
  • TNs provide intrinsic measures of complexity and correlation, enabling better interpretability of model decisions.
Read more
Optimal last-iterate convergence in matrix games with bandit feedback using the log-barrier
Come Fiegel, Pierre Menard, Tadashi Kozuno, Michal Valko, Vianney Perchet
Theory Optimization
  • Introduces a new algorithm achieving ˜O(t−1/4) last-iterate convergence in bandit settings.
  • Utilizes log-barrier regularization and dual-focused analysis for improved convergence rates.
  • Extends results to extensive-form games, maintaining the same convergence rate.
  • Addresses the limitations of previous methods that did not achieve optimal rates.
Read more
Can LLMs Score Medical Diagnoses and Clinical Reasoning as well as Expert Panels?
Amy Rouillard, Sitwala Mundiab, Linda Camarab, Michael Cameron Gramaniec, Ziyaad Dangorc, Ismail Kallad, Shabir A. Madhic, Kajal Morarc, Marlvin T. Ncubec, Haroon Saloojeee, Bruce A. Bassett
Large Language Models NLP Multimodal
  • LLM jury scores are systematically lower than expert clinician panel scores.
  • LLM jury shows better concordance with primary expert panels than human re-scoring panels.
  • The probability of severe errors is lower in LLM jury models compared to human experts.
  • Calibration of LLM jury improves alignment with human expert evaluations.
Read more
Catching Every Ripple: Enhanced Anomaly Awareness via Dynamic Concept Adaptation
Jiaqi Zhu, Shaofeng Cai, Jie Chen, Fang Deng, Beng Chin Ooi, Wenqiao Zhang
Time Series Theory Efficient ML
  • DyMETER integrates parameter shifting and dynamic thresholding for effective online anomaly detection.
  • The framework adapts to new concepts without retraining, enhancing efficiency.
  • Instance-level concept uncertainty is estimated for robust adaptation.
  • Dynamic threshold optimization ensures continuous alignment with evolving data distributions.
Read more
AdaSplash-2: Faster Differentiable Sparse Attention
Nuno Gonçalves, Hugo Pitorro, Vlad Niculae, Edoardo Ponti, Lei Li, Andre Martins, Marcos Treviso
NLP Large Language Models Efficient ML
  • ADASPLASH-2 reduces the number of iterations for computing the normalizer τ in α-entmax attention to 1-2.
  • The method utilizes a histogram-based initialization stored in on-chip SRAM for efficient computation.
  • It outperforms FlashAttention-2 in moderate-to-high sparsity regimes, improving training speed.
  • Empirical results show that ADASPLASH-2 matches or exceeds the performance of softmax attention in various tasks.
Read more
Scouting By Reward: VLM-TO-IRL-Driven Player Selection For Esports
Qing Yan, Wenyu Yang, Yufei Wang, Wenhao Ma, Linchong Hu, Yifei Jin, Anton Dahbura
Reinforcement Learning Multimodal
  • Introduces a novel application of inverse reinforcement learning for style-based scouting in esports.
  • Develops a two-branch architecture that combines telemetry data with tactical commentary for player evaluation.
  • Demonstrates that the proposed system can match expert human analysts in player selection accuracy.
  • Addresses the scalability issues of traditional scouting methods by automating the evaluation process.
Read more
Awakening Dormant Experts: Counterfactual Routing to Mitigate MoE Hallucinations
Wentao Hu, Yanbo Zhai, Xiaohui Hu, Mingkuan Zhao, Shanhong Yu, Xue Liu, Kaidong Yu, Shuangyong Song, Xuelong Li
NLP Large Language Models Efficient ML
  • Identifies the 'Dormant Expert' phenomenon in MoE models due to static Top-k routing.
  • Proposes Counterfactual Routing (CoR) as a training-free framework for expert reallocation.
  • Demonstrates a 3.1% average improvement in factual accuracy on hallucination benchmarks.
  • Maintains the same total activation count during inference, ensuring computational efficiency.
Read more
A KL Lens on Quantization: Fast, Forward-Only Sensitivity for Mixed-Precision SSM-Transformer Models
Jason Kong, Nilesh Prasad Pandey, Flavio Ponzina, Tajana Rosing
NLP Large Language Models Efficient ML
  • Introduces a gradient-free sensitivity analysis framework for hybrid SSM-Transformer models.
  • Demonstrates that KL divergence is a superior metric for quantization sensitivity in language models.
  • Validates the proposed method through extensive experiments and real-world profiling.
  • Achieves significant model compression with minimal accuracy loss, suitable for edge deployment.
Read more
Improving Sparse Autoencoder with Dynamic Attention
Dongsheng Wang, Jinsen Zhang, Dawei Su, Hui Huang
Interpretability Computer Vision NLP
  • Introduction of a transformer-based SAE framework that utilizes cross-attention for coherent concept learning.
  • Development of a sparsemax function that dynamically determines the number of active concepts without requiring hyperparameter tuning.
  • Demonstration of superior reconstruction results and coherent concept capture compared to traditional SAE methods.
  • Validation across various image and text tasks, showcasing the flexibility and efficiency of the proposed approach.
Read more
MAny: Merge Anything for Multimodal Continual Instruction Tuning
Zijian Gao, Wangwang Jia, Xingxing Zhang, Pengfei Qian, Tao Sun, Bo Ding, Yong Dou, Huaimin Wang, Kele Xu
Multimodal Large Language Models Efficient ML
  • Identification of a dual-forgetting phenomenon in MLLMs affecting both perception and reasoning.
  • Introduction of Cross-modal Projection Merging (CPM) for adaptive merging of visual features.
  • Development of Low-rank Parameter Merging (LPM) using Recursive Least Squares for optimal parameter merging.
  • MAAny achieves state-of-the-art performance on UCIT and MLLM-DCL benchmarks without GPU training.
Read more
One-shot learning for the complex dynamical behaviors of weakly nonlinear forced oscillators
Teng Ma, Luca Rosafalco, Wei Cui, Lin Zhao, Attilio Frangi
Theory Efficient ML
  • Introduction of a one-shot learning method for nonlinear frequency-response curves.
  • Development of MEv-SINDy for non-autonomous multi-frequency dynamics.
  • Utilization of Generalized Harmonic Balance for complex response decomposition.
  • Validation on MEMS applications showing accurate predictions across excitation levels.
Read more
Pareto-Optimal Offline Reinforcement Learning via Smooth Tchebysheff Scalarization
Aadyot Bhatnagar, Peter Mørch Groth, Ali Madani
Reinforcement Learning Optimization Large Language Models
  • Introduces STOMP, a novel offline RL algorithm for multi-objective optimization.
  • Utilizes smooth Tchebysheff scalarization to effectively capture non-convex regions of the Pareto front.
  • Demonstrates superior performance over existing methods in protein engineering tasks.
  • Addresses the limitations of linear reward scalarization in multi-objective RL.
Read more
CLion: Efficient Cautious Lion Optimizer with Enhanced Generalization
Feihu Huang, Guanyi Zhang, Songcan Chen
Optimization Theory Efficient ML
  • The generalization error of the Lion optimizer is established as O(1/NτT).
  • CLion optimizer improves generalization error to O(1/N).
  • CLion demonstrates a fast convergence rate of O(√d/T^(1/4)).
  • The study provides a rigorous analysis of the generalization properties of learning-based optimizers.
Read more
Zeroth-Order Optimization at the Edge of Stability
Minhak Song, Liang Zhang, Bingcong Li, Niao He, Michael Muehlebach, Sewoong Oh
Optimization Theory
  • Introduces a mean-square linear stability theory for Zeroth-Order optimization methods.
  • Establishes that ZO methods' stability is influenced by the entire Hessian spectrum, unlike First-Order methods.
  • Derives tractable stability bounds based on the largest eigenvalue and Hessian trace.
  • Empirical evidence shows ZO methods operate at the edge of stability across various deep learning tasks.
Read more
Tight Sample Complexity Bounds for Best-Arm Identification Under Bounded Systematic Bias
Tianhao Qian
Theory Optimization Robotics
  • Introduces a localized Best-Arm Identification framework for node expansion under bounded systematic bias.
  • Establishes tight sample complexity bounds for safe node elimination, highlighting the importance of the empirical reward gap.
  • Presents the PAC-MCTS algorithm as a practical implementation of the theoretical findings.
  • Demonstrates through experiments that the proposed method effectively preserves optimal paths while managing computational costs.
Read more
Parameter Importance is Not Static: Evolving Parameter Isolation for Supervised Fine-Tuning
Zekai Lin, Chao Xue, Di Liang, Xingsheng Han, Peiyang Liu, Xianjie Wu, Lei Jiang, Yu Lu, Haibo Shi, Shuang Liang, Minlong Peng
NLP Large Language Models Optimization
  • Parameter importance in supervised fine-tuning is dynamic, not static.
  • Evolving Parameter Isolation (EPI) adapts isolation masks based on online gradient estimates.
  • EPI improves stability and generalization in multi-task learning scenarios.
  • The framework effectively balances the retention of established knowledge with the acquisition of new capabilities.
Read more
LongCoT: Benchmarking Long-Horizon Chain-of-Thought Reasoning
Sumeet Ramesh Motwani, Daniel Nichols, Charles London, Peggy Li, Fabio Pizzati, Acer Blake, Hasan Hammoud, Tavish McDonald, Akshat Naik, Alesia Ivanova, Vignesh Baskaran, Ivan Laptev, Ruben Glatt, Tal Ben-Nun, Philip Torr, Natasha Jaques, Ameya Prabhu, Brian Bartoldson, Bhavya Kailkhura, Christian Schroeder de Witt
Large Language Models NLP Theory
  • LongCoT is a novel benchmark for evaluating long-horizon reasoning in language models.
  • The benchmark consists of 2,500 expert-designed problems across multiple domains.
  • Current top models achieve less than 10% accuracy on LongCoT, highlighting significant reasoning limitations.
  • The problems require navigating complex interdependencies, emphasizing the need for planning and error management.
Read more
Drowsiness-Aware Adaptive Autonomous Braking System based on Deep Reinforcement Learning for Enhanced Road Safety
Hossem Eddine Hafidi, Elisabetta De Giovanni, Teodoro Montanaro, Ilaria Sergi, Massimo De Vittorio, Luigi Patrono
Reinforcement Learning Robotics Time Series
  • Integration of real-time drowsiness detection into an autonomous braking system.
  • Utilization of ECG signals for accurate drowsiness monitoring.
  • Development of a Double Dual Deep Q-Network (DD-DQN) for adaptive braking policies.
  • Achieved a 99.99% success rate in avoiding accidents in both drowsy and non-drowsy scenarios.
Read more
Enhancing LLM-based Search Agents via Contribution Weighted Group Relative Policy Optimization
Junzhe Wang, Zhiheng Xi, Yajie Yang, Hao Luo, Shihan Dou, Tao Gui, Qi Zhang
NLP Large Language Models Reinforcement Learning Optimization
  • Introduction of Contribution-Weighted GRPO (CW-GRPO) framework for search agents.
  • Reframing process supervision as advantage reallocation based on round contributions.
  • Demonstrated significant performance improvements over standard GRPO.
  • Empirical evidence shows concentrated contributions in informative search rounds.
Read more
Unsupervised domain transfer: Overcoming signal degradation in sleep monitoring by increasing scoring realism
Mohammad Ahangarkiasari, Andreas Tind Damgaard, Casper Haurum, Kaare B. Mikkelsen
Time Series
  • The study investigates the potential of unsupervised domain transfer for sleep monitoring amidst signal degradation.
  • A discriminator-guided approach is proposed to enhance the realism of hypnograms, which can improve scoring accuracy.
  • The unsupervised method shows performance improvements in various signal distortion scenarios without decreasing overall performance.
  • Real-life application of the method revealed limited benefits, indicating the need for further refinement.
Read more
Assessing the Performance-Efficiency Trade-off of Foundation Models in Probabilistic Electricity Price Forecasting
Jan Niklas Lettner, Hadeer El Ashhab, Veit Hagenmeyer, Benjamin Schäfer
Time Series
  • TSFMs generally outperform task-specific models in probabilistic electricity price forecasting.
  • Well-configured task-specific models can achieve performance close to or surpassing TSFMs under certain conditions.
  • The study emphasizes the importance of balancing computational efficiency with forecasting accuracy.
  • Probabilistic forecasts are crucial for managing risks in electricity markets with high renewable energy integration.
Read more
Chain of Uncertain Rewards with Large Language Models for Reinforcement Learning
Shentong Mo
Reinforcement Learning Large Language Models Optimization
  • Introduction of CoUR framework for efficient reward function design in RL.
  • Integration of code uncertainty quantification to streamline reward component reuse.
  • Utilization of Bayesian optimization for independent optimization of reward terms.
  • Extensive evaluation showing CoUR outperforms traditional methods in performance and cost.
Read more
Stability and Generalization in Looped Transformers
Asher Labovich
Theory Large Language Models NLP
  • Introduces a fixed-point framework to analyze stability in looped transformers.
  • Proves that recall and outer normalization are crucial for achieving meaningful predictions.
  • Empirical results confirm that performance aligns with theoretical predictions across various tasks.
  • Presents internal recall as a novel variant that enhances performance under specific conditions.
Read more
Minimax Optimality and Spectral Routing for Majority-Vote Ensembles under Markov Dependence
Ibne Farabi Shihab, Sanjeda Akter, Anuj Sharma
Theory Reinforcement Learning Time Series
  • Establishes a minimax lower bound for classification risk under Markov dependence.
  • Demonstrates that uniform bagging is suboptimal, with a significant risk gap.
  • Proposes adaptive spectral routing to achieve optimal performance in Markov settings.
  • Validates theoretical predictions through extensive experiments on various datasets.
Read more
xFODE: An Explainable Fuzzy Additive ODE Framework for System Identification
Ertugrul Kececi, Tufan Kumbasar
Interpretability Time Series Theory
  • xFODE enhances interpretability in system identification by defining states incrementally based on measurable outputs.
  • The framework utilizes Fuzzy Additive Models to approximate state derivatives, allowing for clearer understanding of input contributions.
  • Partitioning Strategies are introduced to simplify the antecedent space, improving interpretability and reducing inference complexity.
  • xFODE achieves accuracy comparable to existing models like NODE, FODE, and NLARX while providing interpretable insights.
Read more
Adaptive Memory Crystallization for Autonomous AI Agent Learning in Dynamic Environments
Rajat Khanda, Mohammad Baqar, Sambuddha Chakrabarti, Satyasaran Changdar
Reinforcement Learning Robotics Theory
  • Introduction of Adaptive Memory Crystallization (AMC) for continual reinforcement learning.
  • Development of a three-phase memory hierarchy (Liquid, Glass, Crystal) to manage memory stability and plasticity.
  • Rigorous mathematical proofs establishing the convergence and performance guarantees of the proposed SDE.
  • Empirical results show substantial improvements in learning efficiency and memory management.
Read more
Expressivity of Transformers: A Tropical Geometry Perspective
Ye Su, Yong Liu
Theory
  • Introduces a tropical geometry framework to analyze transformers' expressivity.
  • Establishes that self-attention corresponds to a Power Voronoi Diagram in the zero-temperature limit.
  • Demonstrates that Multi-Head Self-Attention expands complexity to O(N H).
  • Derives the first tight bounds on the number of linear regions in transformers as Θ(N dmodel L).
Read more
Enhancing Confidence Estimation in Telco LLMs via Twin-Pass CoT-Ensembling
Anton Saenko, Pranshav Gajjar, Abiodun Ganiyu, Vijay K. Shah
NLP Large Language Models
  • Identifies systematic overconfidence in LLM-generated confidence scores in telecommunications.
  • Proposes a Twin-Pass CoT-Ensembling method to improve confidence estimation.
  • Achieves up to 88% reduction in Expected Calibration Error (ECE) across benchmarks.
  • Provides empirically validated confidence thresholds and recommendations for telecom applications.
Read more
Wasserstein Formulation of Reinforcement Learning. An Optimal Transport Perspective on Policy Optimization
Mathias Dus
Reinforcement Learning Optimization Theory
  • Introduces a geometric framework for RL using Wasserstein space.
  • Establishes rigorous existence guarantees for stationary distributions.
  • Utilizes Otto's calculus for second-order analysis of policy optimization.
  • Demonstrates scalability of the method to high-dimensional problems.
Read more
Some Theoretical Limitations of t-SNE
Rupert Li, Elchanan Mossel
Theory
  • t-SNE can lose important data features during dimensionality reduction.
  • In high-dimensional spaces, t-SNE may map distinct points to the same location in lower dimensions.
  • The paper provides mathematical propositions demonstrating the limitations of t-SNE in preserving data structure.
  • The findings suggest that t-SNE may not be appropriate for all datasets, particularly those with high dimensionality.
Read more
MixAtlas: Uncertainty-aware Data Mixture Optimization for Multimodal LLM Midtraining
Bingbing Wen, Sirajul Salekin, Feiyang Kang, Bill Howe, Lucy Lu Wang, Javier Movellan, Manjot Bilkhu
Multimodal Optimization Large Language Models
  • MixAtlas provides a two-axis data decomposition for interpretable multimodal mixture optimization.
  • The method utilizes small proxy models and Gaussian-process surrogates for efficient mixture search.
  • Empirical results show significant performance improvements and faster convergence compared to existing baselines.
  • Recipes developed on smaller models are transferable to larger models, enhancing practical optimization.
Read more
Quantization of Spiking Neural Networks Beyond Accuracy
Evan Gibson Smith, Jacob Whitehill, Fatemeh Ganji
Efficient ML
  • EMD is proposed as a diagnostic metric for evaluating firing distribution divergence in quantized SNNs.
  • Quantization methods significantly affect firing distributions, even when accuracy is preserved.
  • Learned quantization (e.g., LQ-Net) maintains firing behavior more effectively than uniform quantization.
  • The study highlights the importance of considering firing dynamics in the deployment of quantized SNNs.
Read more
First-See-Then-Design: A Multi-Stakeholder View for Optimal Performance-Fairness Trade-Offs
Kavya Gupta, Nektarios Kalampalikis, Christoph Heitz, Isabel Valera
Theory Optimization
  • Introduces a multi-stakeholder framework for fair algorithmic decision-making.
  • Shifts focus from prediction-centric fairness to utility-based fairness.
  • Utilizes post-hoc multi-objective optimization to explore performance-fairness trade-offs.
  • Demonstrates that stochastic policies can yield better outcomes than deterministic ones.
Read more
Calibrate-Then-Delegate: Safety Monitoring with Risk and Budget Guarantees via Model Cascades
Edoardo Pona, Milad Kazemi, Mehran Hosseini, Yali Du, David Watson, Osvaldo Simeone, Nicola Paoletti
NLP Large Language Models Theory
  • CTD introduces a model-cascade approach with finite-sample guarantees on delegation rate and safety performance.
  • The delegation value (DV) probe provides a more targeted delegation signal compared to traditional uncertainty measures.
  • CTD consistently outperforms existing methods in safety monitoring across various budget levels.
  • The method adapts budget allocation based on input difficulty, preventing harmful over-delegation.
Read more
How Embeddings Shape Graph Neural Networks: Classical vs Quantum-Oriented Node Representations
Nouhaila Innan, Antonello Rosato, Alberto Marchisio, Muhammad Shafique
Graph Learning
  • Establishes a unified framework for evaluating node embeddings in GNNs.
  • Compares classical and quantum-oriented embeddings under controlled conditions.
  • Demonstrates that quantum embeddings outperform classical ones on structure-driven datasets.
  • Identifies the importance of embedding design in graph-level prediction.
Read more
Hardware-Efficient Neuro-Symbolic Networks with the Exp-Minus-Log Operator
Eymen Ipek
Efficient ML Interpretability Theory
  • Introduction of the Exp-Minus-Log (EML) operator as a unifying primitive for DNNs.
  • Development of a DNN-EML hybrid architecture that enhances interpretability and reduces hardware complexity.
  • Establishment of computational-cost bounds and analysis of inference and training acceleration.
  • Identification of a literature gap in existing neuro-symbolic approaches that do not utilize a single hardware-realizable primitive.
Read more
Bridging MARL to SARL: An Order-Independent Multi-Agent Transformer via Latent Consensus
Zijian Zhao, Jing Gao, Sen Li
Reinforcement Learning Robotics Optimization
  • CMAT bridges MARL and SARL, addressing key challenges in cooperative multi-agent settings.
  • The framework utilizes a Transformer encoder and a hierarchical decision-making mechanism for effective coordination.
  • Simultaneous action generation based on a consensus vector reduces sensitivity to action order.
  • CMAT shows superior performance on benchmark tasks compared to existing methods.
Read more
Robust Ultra Low-Bit Post-Training Quantization via Stable Diagonal Curvature Estimate
Jaemin Kim, Sungkyun Kim, Junyeol Lee, Jiwon Seo
Large Language Models Efficient ML Optimization
  • DASH-Q improves robustness in ultra low-bit quantization by using diagonal Hessian approximations.
  • The framework effectively filters out noise from calibration data, enhancing feature preservation.
  • Achieves significant accuracy improvements over existing PTQ methods, particularly in low-bit regimes.
  • Demonstrates strong performance with minimal calibration data, making it suitable for resource-limited environments.
Read more
ConfLayers: Adaptive Confidence-based Layer Skipping for Self-Speculative Decoding
Walaa Amer, Uday Das, Fadi Kurdahi
NLP Large Language Models Efficient ML
  • ConfLayers introduces a confidence-based mechanism for adaptive layer skipping in self-speculative decoding.
  • The framework is training-free and offers a plug-and-play solution for constructing draft subnetworks.
  • Empirical results show up to 1.4× speedup over standard autoregressive decoding with maintained output quality.
  • ConfLayers consistently outperforms existing heuristic and dynamic skipping methods.
Read more
Multi-User mmWave Beam and Rate Adaptation via Combinatorial Satisficing Bandits
Emre Özyıldırım, Barış Yaycı, Umut Eren Akturk, Cem Tekin
Theory Optimization Efficient ML
  • Introduces SAT-CTS, a lightweight policy for beam and rate adaptation in mmWave systems.
  • Establishes finite-time regret bounds for combinatorial semi-bandits with satisficing objectives.
  • Demonstrates that SAT-CTS effectively reduces satisficing regret while maintaining fairness and throughput.
  • Eliminates the need for explicit channel state information by relying on ACK/NACK feedback.
Read more
Sparse Goodness: How Selective Measurement Transforms Forward-Forward Learning
Kamer Ali Yuksel, Hassan Sawaf
Theory Optimization Efficient ML
  • Introduction of top-k goodness function, significantly outperforming the traditional sum-of-squares method.
  • Development of entmax-weighted energy for adaptive sparse weighting, leading to improved accuracy.
  • Implementation of separate label–feature forwarding (FFCL) enhances performance across all goodness functions.
  • Establishment of a unifying principle that emphasizes the importance of sparsity in goodness functions for FF networks.
Read more
Golden Handcuffs make safer AI agents
Aram Ebtekar, Michael K. Cohen
Reinforcement Learning Theory
  • Introduces the 'Golden Handcuffs' mechanism to enhance safety in AI agents.
  • Expands the reward range to include negative values, promoting risk aversion.
  • Proves that the agent can achieve sublinear regret against the best mentor.
  • Ensures that unsafe actions are only triggered by mentors, not the optimizing policy.
Read more
Blazing the trails before beating the path: Sample-efficient Monte-Carlo planning
Jean-Bastien Grill, Michal Valko, Rémi Munos
Reinforcement Learning Robotics Efficient ML
  • Introduction of TrailBlazer, a sample-efficient Monte-Carlo planning algorithm.
  • Focus on exploring near-optimal states to minimize oracle calls.
  • Theoretical guarantees on sample complexity provided.
  • Comparison with existing algorithms shows significant efficiency improvements.
Read more
Physics-Informed Neural Networks for Methane Sorption: Cross-Gas Transfer Learning, Ensemble Collapse Under Physics Constraints, and Monte Carlo Dropout Uncertainty Quantification
Mohammad Nooraiepour, Zezhang Song, Wei Li, Sarah Perez
Theory Interpretability Efficient ML
  • Introduces a physics-informed transfer learning framework for methane sorption prediction.
  • Achieves a 227% improvement over classical isotherm models in predictive accuracy.
  • Monte Carlo Dropout is identified as the best method for uncertainty quantification.
  • Demonstrates the importance of moisture-volatile interactions in methane sorption.
Read more