AI-generated summaries

Today's ML research,
without the noise.

Daily summaries of the latest machine learning papers from arXiv, processed every 8 hours.

67 Papers today
8h Update frequency
7 Days of history
SafeAdapt: Provably Safe Policy Updates in Deep Reinforcement Learning
Maksim Anisimov, Francesco Belardinelli, Matthew Wicker
Reinforcement Learning Theory Robotics
  • Introduction of the Rashomon set for safe policy updates in RL.
  • Formal guarantees for safety during policy adaptation in non-stationary environments.
  • Empirical validation in grid-world navigation tasks demonstrating superior performance over existing methods.
  • Prevention of catastrophic forgetting of safety constraints during continual learning.
Read more
Using Synthetic Data for Machine Learning-based Childhood Vaccination Prediction in Narok, Kenya
Jimmy Bach, Yang Li, Yaqi Liu, John Sankok, Rose Kimani, Carrie B. Dolan, Julius N. Odhiambo, Haipeng Chen
Generative Models Optimization Theory
  • Identification of children at risk of missing vaccinations can improve healthcare interventions.
  • Synthetic data generation can protect patient privacy without sacrificing predictive accuracy.
  • Machine learning models can effectively predict vaccination risks in low-resource settings.
  • High performance metrics (recall, precision, F1-scores > 90%) were achieved for vaccination predictions.
Read more
Is your algorithm unlearning or untraining?
Eleni Triantafillou, Ahmed Imtiaz Humayun, Monica Ribero, Alexander Matt Turner, Michael C. Mozer, Georgios Kaissis
Theory
  • Distinction between 'Unlearning' and 'Untraining' is crucial for clarity in research.
  • Untraining removes the influence of specific examples, while Unlearning addresses broader distributions.
  • Misuse of the term 'unlearning' can lead to inappropriate metrics and expectations in algorithm evaluation.
  • Clarifying these concepts can accelerate progress in machine unlearning research.
Read more
Efficient RL Training for LLMs with Experience Replay
Charles Arnal, Vivien Cabannes, Taco Cohen, Julia Kempe, Remi Munos
Large Language Models Reinforcement Learning Efficient ML
  • Experience Replay can enhance sample efficiency in LLM post-training, contrary to prevailing beliefs.
  • Theoretical analysis provides a framework for optimizing replay buffer design based on compute efficiency and data diversity.
  • Empirical results show that using a replay buffer can save up to 40% of compute budget while maintaining or improving model accuracy.
  • The study emphasizes the importance of balancing data staleness and diversity for optimal training outcomes.
Read more
Finite-Sample Analysis of Nonlinear Independent Component Analysis: Sample Complexity and Identifiability Bounds
Yuwen Jiang
Theory
  • Establishes the first complete characterization of finite-sample analysis for nonlinear ICA.
  • Identifies sample complexity scaling laws that guide practitioners in determining sample sizes.
  • Introduces a direct relationship between excess risk and identification error, improving convergence rates.
  • Validates theoretical predictions through extensive simulations, confirming scaling laws with high accuracy.
Read more
Temporal Dropout Risk in Learning Analytics: A Harmonized Survival Benchmark Across Dynamic and Early-Window Representations
Rafael da Silva, Jeff Eicher, Gregory Longo
Time Series
  • A harmonized benchmark for dropout risk modeling enhances comparability across different predictive models.
  • Temporal and behavioral signals are more predictive of dropout risk than static demographic factors.
  • Calibration and interpretability are essential for practical applications of dropout risk predictions.
  • Random Survival Forest and Poisson Piecewise-Exponential models show strong performance in their respective arms.
Read more
From Selection to Scheduling: Federated Geometry-Aware Correction Makes Exemplar Replay Work Better under Continual Dynamic Heterogeneity
Zhuang Qi, Ying-Peng Tang, Lei Meng, Guoqing Chao, Lei Wu, Han Yu, Xiangxu Meng
Federated Learning
  • FEAT addresses the limitations of exemplar replay in federated continual learning by focusing on both sample selection and effective utilization.
  • The Geometric Structure Alignment module enhances feature consistency across clients by aligning local representations with shared prototypes.
  • The Energy-based Geometric Correction module mitigates prediction bias towards majority classes, improving sensitivity to minority classes.
  • Experimental results show that FEAT consistently outperforms seven state-of-the-art methods in terms of accuracy across different datasets.
Read more
ECHO: Efficient Chest X-ray Report Generation with One-step Block Diffusion
Lifeng Chen, Tianqi You, Hao Liu, Zhimin Bao, Jile Jiao, Xiao Han, Zhicai Ou, Tao Sun, Xiaofeng Mou, Xiaojie Jin, Yi Xu
Computer Vision NLP Generative Models
  • ECHO achieves one-step-per-block parallel decoding for efficient CXR report generation.
  • Introduces Direct Conditional Distillation (DCD) for improved inference speed with minimal quality loss.
  • Response-Asymmetric Diffusion (RAD) adaptation reduces training complexity significantly.
  • Outperforms existing autoregressive and diffusion-based models in clinical accuracy and speed.
Read more
Tree-of-Evidence: Efficient "System 2" Search for Faithful Multimodal Grounding
Micky C. Nnamdi, Benoit L. Marteau, Yishan Zhong, J. Ben Tamo, May D. Wang
Multimodal Interpretability Optimization
  • Introduces Tree-of-Evidence (ToE) for improved interpretability of multimodal models.
  • Frames interpretability as a discrete optimization problem using a beam search strategy.
  • Maintains high predictive performance while producing compact evidence sets.
  • Demonstrates adaptability in evidence selection based on the context of the data.
Read more
From Dispersion to Attraction: Spectral Dynamics of Hallucination Across Whisper Model Scales
Ivan Viakhirev, Kirill Borodin, Grach Mkrtchian
Audio & Speech Theory Interpretability
  • Introduction of the Spectral Sensitivity Theorem to explain hallucinations in ASR models.
  • Identification of two regimes: Structural Disintegration in smaller models and Compression-Seeking Attractor in larger models.
  • Validation of theoretical predictions through eigenspectral analysis of Whisper models under adversarial conditions.
  • Demonstration of how model scaling impacts the dynamics of signal propagation and hallucination occurrences.
Read more
MIPT-SSM: Scaling Language Models with O(1) Inference Cache via Phase Transitions
Yasong Fan
NLP Large Language Models Theory
  • Introduces a learned measurement rate to navigate between wave and particle phases in sequence modeling.
  • Demonstrates a significant reduction in memory usage while improving accuracy over traditional Transformer models.
  • Establishes a formal proof of the incompatibility between norm-preservation and selective forgetting in linear operators.
  • Empirical validation across multiple tasks including text classification and language modeling.
Read more
Identification and Anonymization of Named Entities in Unstructured Information Sources for Use in Social Engineering Detection
Carlos Jimeno Miguel, Raul Orduna, Francesco Zola
NLP Audio & Speech Multimodal
  • Development of a system for collecting and processing unstructured data from Telegram.
  • Implementation of advanced speech-to-text models for audio data transcription.
  • Evaluation of NER solutions, highlighting the effectiveness of transformer-based architectures.
  • Introduction of anonymization metrics to balance data protection and analytical utility.
Read more
Inside-Out: Measuring Generalization in Vision Transformers Through Inner Workings
Yunxiang Peng, Mengmeng Ma, Ziyu Yao, Xi Peng
Computer Vision Interpretability Theory
  • Introduces two new metrics for evaluating model generalization based on internal mechanisms.
  • Dependency Depth Bias (DDB) measures reliance on deep versus shallow features before deployment.
  • Circuit Shift Score (CSS) predicts model performance under distribution shifts after deployment.
  • Both metrics show improved correlation with OOD performance, outperforming existing proxies.
Read more
On the Spectral Geometry of Cross-Modal Representations: A Functional Map Diagnostic for Multimodal Alignment
Krisanu Sarkar
Multimodal
  • First application of functional maps to multimodal neural representation alignment.
  • Evidence that independently pretrained vision and language encoders develop similar spectral complexity.
  • Identification of the spectral complexity–orientation gap in cross-modal representations.
  • Introduction of three quantitative diagnostics for assessing representation compatibility.
Read more
On the Role of DAG topology in Energy-Aware Cloud Scheduling : A GNN-Based Deep Reinforcement Learning Approach
Anas Hattay, Fred Ngole Mboula, Eric Gascard, Zakaria Yahoun
Reinforcement Learning Graph Learning Optimization
  • Introduces a GNN-based DRL approach for scheduling workflows in cloud environments.
  • Identifies OOD conditions that cause performance degradation in GNN-based schedulers.
  • Analyzes the impact of structural mismatches on message passing and policy generalization.
  • Highlights the need for robust representations to improve scheduling performance under distribution shifts.
Read more
Conservation Law Breaking at the Edge of Stability: A Spectral Theory of Non-Convex Neural Network Optimization
Daniel Nobrega Medeiros
Optimization Theory
  • Gradient descent on L-layer ReLU networks preserves L−1 conservation laws under continuous flow.
  • Discrete gradient descent breaks these conservation laws, leading to a drift characterized by a power law.
  • A closed-form spectral crossover formula for drift is derived and validated across multiple architectures.
  • Cross-entropy loss induces exponential spectral compression in the Hessian, independent of dataset size.
Read more
Automated Batch Distillation Process Simulation for a Large Hybrid Dataset for Deep Anomaly Detection
Jennifer Werner, Justus Arweiler, Indra Jungjohann, Jochen Schmid, Fabian Jirasek, Hans Hasse, Michael Bortz
Time Series
  • Creation of a large hybrid dataset combining experimental and simulated data for batch distillation.
  • Development of an automated Python-based process simulator for generating consistent simulation data.
  • The hybrid dataset is openly released and includes rich metadata and anomaly annotations.
  • Addresses the challenges of obtaining annotated data from real chemical processes.
Read more
Act or Escalate? Evaluating Escalation Behavior in Automation with Language Models
Matthew DosSantos DiSorbo, Harang Ju
NLP Large Language Models Theory
  • LLMs exhibit miscalibrated self-assessments, leading to inconsistent escalation behavior.
  • Escalation thresholds vary significantly across models and are not predicted by architecture or scale.
  • Interventions such as supervised fine-tuning on chain-of-thought targets yield robust decision-making policies.
  • Effective automation requires LLMs to explicitly reason about uncertainty and decision costs.
Read more
Are Independently Estimated View Uncertainties Comparable? Unified Routing for Trusted Multi-View Classification
Yilin Zhang, Cai Xu, Haishun Chen, Ziyu Guan, Wei Zhao
Multimodal
  • Identifies the fragility of the assumption that evidence from different views is comparable.
  • Proposes TMUR, which decouples evidence extraction from fusion arbitration.
  • Introduces a unified router that generates sample-level expert weights based on global context.
  • Demonstrates improved classification performance and reliability through extensive experiments.
Read more
Flow Learners for PDEs: Toward a Physics-to-Physics Paradigm for Scientific Computing
Yilong Dai, Shengyu Chen, Xiaowei Jia, Runlong Yu
Theory Generative Models Optimization
  • Current learned PDE solvers often rely on state prediction, which is inadequate for complex scientific problems.
  • Flow learners parameterize transport vector fields, providing a more accurate representation of PDE dynamics.
  • The proposed approach enhances uncertainty quantification and supports continuous-time predictions.
  • A shift from regression-based models to transport-based learning is necessary for effective PDE solving.
Read more
Efficient Hierarchical Implicit Flow Q-learning for Offline Goal-conditioned Reinforcement Learning
Zhiqiang Dong, Teng Pang, Rongjian Xu, Guoqiang Wu
Reinforcement Learning Generative Models Robotics
  • Introduction of Hierarchical Implicit Flow Q-Learning (HIFQL) for offline GCRL.
  • Utilization of a goal-conditioned mean flow policy to enhance hierarchical policy expressiveness.
  • Incorporation of LeJEPA loss for improved goal representation and generalization.
  • Demonstrated strong performance on OGBench benchmark for both state-based and pixel-based tasks.
Read more
Adaptive Candidate Point Thompson Sampling for High-Dimensional Bayesian Optimization
Donney Fan, Geoff Pleiss
Optimization
  • Introduction of Adaptive Candidate Thompson Sampling (ACTS) to improve Bayesian optimization in high dimensions.
  • ACTS adaptively reduces the search space by generating candidate points in gradient-aligned subspaces.
  • Demonstrated significant performance improvements over traditional TS methods in both synthetic and real-world scenarios.
  • Maintains global consistency, ensuring effective exploration of the optimization landscape.
Read more
SCOT: Multi-Source Cross-City Transfer with Optimal-Transport Soft-Correspondence Objective
Yuyao Wang, Min Yang, Meng Chen, Weiming Huang, Yongshun Gong
Graph Learning Optimization Theory
  • SCOT addresses the central challenge of establishing explicit soft correspondences in cross-city transfer learning.
  • The framework utilizes Sinkhorn-based entropic optimal transport to manage unequal region partitions effectively.
  • An OT-weighted contrastive objective sharpens semantic separability and enhances transferability of learned embeddings.
  • SCOT shows consistent improvements over strong baselines in real-world applications, demonstrating robustness under data heterogeneity.
Read more
Prediction Arena: Benchmarking AI Models on Real-World Prediction Markets
Jaden Zhang, Gardenia Liu, Oliver Johansson, Hileamlak Yitayew, Kamryn Ohly, Grace Li
Reinforcement Learning Theory Optimization
  • Prediction Arena benchmarks AI models in real prediction markets with actual capital.
  • Cohort 1 models showed significant performance differences between platforms, with Kalshi yielding worse returns than Polymarket.
  • The study identifies key drivers of performance, including initial prediction accuracy and the ability to capitalize on correct predictions.
  • Computational efficiency does not correlate with performance, challenging assumptions about model complexity.
Read more
Hierarchical Kernel Transformer: Multi-Scale Attention with an Information-Theoretic Approximation Analysis
Giansalvo Cirrincione
Theory Efficient ML NLP
  • HKT introduces a multi-scale attention mechanism that processes sequences at different resolution levels.
  • The model captures both local and long-range dependencies while maintaining lower computational costs compared to traditional attention.
  • Theoretical contributions include a positive semidefinite kernel definition and a unique decomposition of the asymmetric score matrix.
  • Empirical results show consistent performance improvements across multiple tasks, including synthetic ListOps and CIFAR-10.
Read more
AlphaLab: Autonomous Multi-Agent Research Across Optimization Domains with Frontier LLMs
Brendan R. Hogan, Xiwen Chen, James T. Wilson, Kashif Rasul, Adel Boyarsky, Thomas Kamei, Anderson Schneider, Yuriy Nevmyvaka
Large Language Models Optimization Time Series
  • ALPHALAB automates the full experimental cycle in quantitative domains using frontier LLMs.
  • The system operates through three phases: exploration, evaluation framework construction, and large-scale experimentation.
  • ALPHALAB demonstrates significant performance improvements in CUDA kernel optimization, LLM pretraining, and traffic forecasting.
  • The use of multiple LLMs allows for diverse solution discovery, enhancing the research process.
Read more
EngageTriBoost: Predictive Modeling of User Engagement in Digital Mental Health Intervention Using Explainable Machine Learning
Ha Na Cho, Daniel Eisenberg, Cheryl King, Kai Zheng
Interpretability
  • EngageTriBoost (ETB) is an explainable ensemble machine learning framework for predicting user engagement in DMHI.
  • ETB achieved up to 84% accuracy in predicting user message posting, outperforming individual models.
  • The study utilized Shapley Additive Explanations (SHAP) for interpretability, revealing key behavioral and demographic factors affecting engagement.
  • The framework emphasizes the need for understanding multi-level user engagement rather than treating it as a static binary outcome.
Read more
Robust Reasoning Benchmark
Pavel Golikov, Evgenii Opryshko, Gennady Pekhimenko, Mark C. Jeffrey
NLP Large Language Models Theory
  • Introduction of a perturbation pipeline with 14 deterministic transformations for evaluating LLM robustness.
  • Demonstration of significant accuracy degradation in open-weight models under perturbations.
  • Identification of Intra-Query Attention Dilution, where prior reasoning steps negatively impact subsequent tasks.
  • Call for future LLM architectures to integrate mechanisms for contextual resets to improve reasoning reliability.
Read more
Uncertainty-Aware Transformers: Conformal Prediction for Language Models
Abhiram Vellore, Niraj K. Jha
NLP Large Language Models Interpretability
  • Introduction of CONFIDE, a conformal prediction framework for transformer models.
  • Achieves up to 4.09% improvement in test accuracy and higher correct efficiency over existing methods.
  • Demonstrates better calibration and semantic representation in early and intermediate transformer layers.
  • Provides instance-level explanations for predictions, enhancing interpretability.
Read more
Distributed Online Convex Optimization with Compressed Communication: Optimal Regret and Applications
Sifan Yang, Dan-Yue Li, Lijun Zhang
Optimization Theory Efficient ML
  • Establishes lower bounds for regret in D-OCO with compressed communication.
  • Introduces the D-FTFCL algorithm that optimally manages compression and projection errors.
  • Achieves optimal regret bounds for both convex and strongly convex loss functions.
  • Extends methods to offline stochastic optimization with domain constraints.
Read more
Dead Weights, Live Signals: Feedforward Graphs of Frozen Language Models
Marcus Armstrong, Navid Ayoobi, Arjun Mukherjee
NLP Large Language Models Efficient ML
  • Introduces a feedforward graph architecture using frozen LLMs as nodes communicating through a shared latent space.
  • Achieves strong benchmark performance, outperforming the best single model and parameter-matched classifiers.
  • Demonstrates tractable gradient flow through multiple frozen model boundaries.
  • Emergent selective routing behavior is observed in the output node without explicit supervision.
Read more
Introducing Echo Networks for Computational Neuroevolution
Christian Kroos, Fabian Küch
Audio & Speech Efficient ML Time Series
  • Introduction of Echo Networks as a new type of recurrent neural network for neuroevolution.
  • Echo Networks utilize a connection matrix for representing topology and weights, allowing for flexible architecture.
  • Demonstrated effectiveness in classifying electrocardiography signals with minimal network size.
  • Enhanced systematicity in mutation and recombination processes compared to traditional methods.
Read more
MolPaQ: Modular Quantum-Classical Patch Learning for Interpretable Molecular Generation
Syed Rameez Naqvi, Lu Peng
Generative Models Graph Learning Interpretability
  • MOLPAQ integrates quantum circuits as modular patch generators within a classical conditioning framework.
  • The architecture enforces chemical realism through a constraint-aware molecular assembly process.
  • MOLPAQ achieves high validity, novelty, and diversity in generated molecules.
  • The framework allows for improved property control and interpretability in molecular generation.
Read more
HiFloat4 Format for Language Model Pre-training on Ascend NPUs
Mehran Taghian, Yunke Peng, Xing Huang, Yao Wang, Yaoyuan Wang, Wei Guo, Yuanyong Luo, Tianchi Hu, Junsong Wang, Xin Wang, Hu Liu, Yu Cheng, Ziwei Yu, Hongliang Li, Mehdi Rahimifar, Lei Yan, Xuefei Wang, Zhuang Ma, Lei Liu, Hui Yu, Anandharaju Durai Raju, Hoang Le, Hei Yi Mak, Tanzila Rahman, Shadan Golestan
Large Language Models Efficient ML
  • First study on low-precision LLM pre-training on energy-efficient NPU accelerators.
  • HiFloat4 format achieves lower relative loss (≈1.0%) compared to MXFP4 (≈1.5%).
  • HiF4 training requires fewer stabilization techniques than MXFP4.
  • Stable training of both dense and Mixture-of-Experts LLM architectures is demonstrated.
Read more
LLM-Generated Fault Scenarios for Evaluating Perception-Driven Lane Following in Autonomous Edge Systems
Faezeh Pasandideh, Achim Rettberg
Computer Vision Generative Models Robotics
  • Introduction of a decoupled offline–online fault injection framework for autonomous systems.
  • Utilization of LLMs for generating fault scenarios and LDMs for synthesizing visual degradations.
  • Real-time fault condition assessment on edge devices without heavy computational load.
  • Significant robustness degradation observed under various fault scenarios, emphasizing the need for comprehensive testing.
Read more
Offline Local Search for Online Stochastic Bandits
Gerdus Benadè, Rathish Das, Thomas Lavastida
Optimization Theory
  • Introduces a framework for converting offline local search algorithms into online stochastic bandit algorithms.
  • Achieves O(log³ T) regret, improving upon existing methods that yield polynomial regret.
  • Applies the framework to three combinatorial optimization problems, showcasing its versatility.
  • Establishes the conditions under which local search algorithms can guarantee γ-regret.
Read more
GNN-as-Judge: Unleashing the Power of LLMs for Graph Learning with GNN Feedback
Ruiyao Xu, Kaize Ding
Large Language Models Graph Learning
  • Introduces GNN-as-Judge framework to improve LLM performance on TAGs in low-resource settings.
  • Addresses challenges of generating reliable pseudo labels and mitigating label noise during fine-tuning.
  • Utilizes GNNs to provide structural insights for selecting influential unlabeled nodes.
  • Demonstrates significant performance improvements over existing methods in few-shot semi-supervised learning.
Read more
Practical Bayesian Inference for Speech SNNs: Uncertainty and Loss-Landscape Smoothing
Yesmine Abdennadher, Philip N. Garner
Audio & Speech Theory Optimization
  • Introduces IVON as a Bayesian learning method for SNNs, addressing weight uncertainty.
  • Demonstrates that Bayesian learning can smooth the irregular predictive landscape of SNNs.
  • Shows improved performance on speech recognition tasks using IVON compared to deterministic methods.
  • Provides evidence through predictive, calibration, and local loss geometry analyses.
Read more
GAN-based Domain Adaptation for Image-aware Layout Generation in Advertising Poster Design
Chenchen Xu, Min Zhou, Tiezheng Ge, Weiwei Xu
Generative Models Computer Vision
  • Introduction of the CGL-Dataset with over 60,000 paired posters and 121,000 product images.
  • Development of two GAN-based models: CGL-GAN and PDA-GAN, to address domain adaptation challenges.
  • PDA-GAN employs a pixel-level discriminator for improved layout generation based on image content.
  • Three novel content-aware metrics are proposed for evaluating layout generation quality.
Read more
Ranked Activation Shift for Post-Hoc Out-of-Distribution Detection
Gianluca Guglielmo, Marc Masana
Computer Vision
  • Introduces RAS, a hyperparameter-free method for OoD detection that enhances performance without requiring retraining.
  • Identifies the limitations of existing scaling-based methods and the impact of unrectified activations on performance.
  • Demonstrates consistent performance across different datasets and architectures, preserving in-distribution accuracy.
  • Analyzes the contributions of both inhibitory and excitatory activation shifts to the effectiveness of OoD detection.
Read more
EgoEverything: A Benchmark for Human Behavior Inspired Long Context Egocentric Video Understanding in AR Environment
Qiance Tang, Ziqi Wang, Jieyu Lin, Ziyun Li, Barbara De Salvo, Sai Qian Zhang
Computer Vision Multimodal
  • EgoEverything incorporates human attention into question generation for long-context egocentric video understanding.
  • The benchmark features over 5,000 question-answer pairs across 100+ hours of video, reflecting real-world AR interactions.
  • A novel Visual Question Answering pipeline and attention-inspired sampling strategy are utilized to generate questions.
  • Evaluation reveals that current vision-language models perform poorly on EgoEverything, highlighting their limitations in real-life scenarios.
Read more
Adversarial Label Invariant Graph Data Augmentations for Out-of-Distribution Generalization
Simon Zhang, Ryan P. DeMilt, Kun Jin, Cathy H. Xia
Graph Learning Optimization Theory
  • Introduction of RIA, a method for improving OoD generalization under covariate shift.
  • Formulation of adversarial label invariant data augmentations to create diverse training environments.
  • Development of an alternating gradient descent-ascent algorithm for optimization.
  • Extensive experimental validation showing superior performance over existing OoD methods.
Read more
Mathematical analysis of one-layer neural network with fixed biases, a new activation function and other observations
Fabricio Macià, Shu Nakamura
Theory
  • Rigorous proof of convergence for one-hidden-layer neural networks with fixed biases using gradient descent.
  • Introduction of a new activation function, FReX, which maintains convergence properties.
  • Establishment of the spectral bias property for the learning process.
  • Discussion on the representability of functions in both continuous and discrete models.
Read more
Loom: A Scalable Analytical Neural Computer Architecture
Mehmet Kerem Turkcan
Theory Efficient ML Interpretability
  • Loom implements a 22-opcode instruction set in 8 transformer layers, optimizing execution efficiency.
  • The opcode-as-operand-routing technique reduces the complexity of execution layers, enabling a more compact architecture.
  • The introduction of the STORE instruction significantly decreases the size of compiled programs, enhancing performance.
  • A C compiler facilitates the translation of C programs to Loom's ISA, supporting both offline and in-browser execution.
Read more
Structured Exploration and Exploitation of Label Functions for Automated Data Annotation
Phong Lam, Ha-Linh Nguyen, Thu-Trang Nguyen, Son Nguyen, Hieu Dinh Vo
NLP Efficient ML Theory
  • EXPONA introduces a two-phase process for LF generation that balances diversity and reliability.
  • The framework systematically explores multi-level label functions to improve coverage and quality.
  • EXPONA achieved up to 98.9% label coverage and improved weak label quality by up to 87%.
  • The method resulted in substantial downstream performance gains, with improvements of up to 46% in weighted F1 scores.
Read more
Bit-by-Bit: Progressive QAT Strategy with Outlier Channel Splitting for Stable Low-Bit LLMs
Binxing Xu, Hao Gu, Lujun Li, Hao Wang, Bei Liu, Jiacheng Liu, Qiyuan Zhu, Xintong Yang, Chao Li, Sirui Han, Yike Guo
Large Language Models Efficient ML Optimization
  • Introduces a progressive QAT framework that enhances stability during low-bit training.
  • Employs outlier channel splitting to mitigate quantization errors effectively.
  • Achieves significant speed improvements with custom operators for low-bit configurations.
  • Demonstrates superior performance on LLaMA-2/3 compared to existing QAT baselines.
Read more
Feature-Label Modal Alignment for Robust Partial Multi-Label Learning
Yu Chen, Weijun Lv, Yue Huang, Xiaozhao Fang, Jie Wen, Yong Xu, Guanbin Li
Multimodal
  • Introduction of modal alignment to improve feature-label consistency in PML.
  • Development of a low-rank orthogonal decomposition method for robust pseudo-label generation.
  • Implementation of a multi-peak class prototype learning mechanism to enhance discriminability.
  • Demonstrated superior performance in classification accuracy and noise robustness compared to existing methods.
Read more
Stability Enhanced Gaussian Process Variational Autoencoders
Carl R. Richardson, Jichen Zhang, Ethan King, Ján Drgoňa
Generative Models Robotics Theory
  • Introduction of SEGP-VAE for training LTI systems using video data.
  • Derivation of mean and covariance functions from LTI system definitions.
  • Unconstrained parametrization prevents numerical issues during training.
  • Demonstrated effectiveness on spiraling particle video dataset.
Read more
SPAMoE: Spectrum-Aware Hybrid Operator Framework for Full-Waveform Inversion
Zhenyu Wang, Peiyuan Li, Yongxiang Shi, Ruoyu Wu, Chenfei Liao, Lei Zhang
Optimization
  • Introduction of SPAMoE, a spectrum-aware framework for FWI.
  • Utilization of a Spectral-Preserving DINO Encoder to maintain frequency balance.
  • Implementation of an Adaptive Spectral Mixture-of-Experts for dynamic frequency band allocation.
  • Significant performance improvement on OpenFWI benchmark datasets.
Read more
Cluster Attention for Graph Machine Learning
Oleg Platonov, Liudmila Prokhorenkova
Graph Learning
  • Introduces Cluster Attention (CLATT) to enhance graph learning models.
  • CLATT allows nodes to attend to all nodes within their clusters, improving long-range dependency capture.
  • Integrating CLATT with MPNNs and Graph Transformers leads to significant performance improvements.
  • The approach leverages community detection algorithms for effective graph partitioning.
Read more
How does Chain of Thought decompose complex tasks?
Amrut Nadgir, Vijay Balasubramanian, Pratik Chaudhari
NLP Large Language Models Theory
  • Classification error in LLMs scales as a power law with the number of classes.
  • Decomposing tasks into smaller classification problems can significantly reduce prediction error.
  • There exists an optimal degree of decomposition that minimizes error, beyond which additional reasoning depth is counterproductive.
  • The study formalizes reasoning as a structured decomposition of classification tasks, explaining empirical observations in LLM performance.
Read more
Toward World Models for Epidemiology
Zeeshan Memon, Yiqi Su, Christo Kurisummoottil Thomas, Walid Saad, Liang Zhao, Naren Ramakrishnan
Time Series Theory Optimization
  • Introduces a formal framework for epidemiological world models.
  • Reframes epidemic decision-making to incorporate latent states and adaptive behaviors.
  • Demonstrates the necessity of world models through empirical case studies.
  • Highlights the limitations of traditional epidemiological models in capturing dynamic behaviors.
Read more
Event-Driven Temporal Graph Networks for Asynchronous Multi-Agent Cyber Defense in NetForge_RL
Igor Jankowski
Reinforcement Learning Graph Learning Time Series
  • Introduction of NetForge_RL, a continuous-time POSMDP simulator for cyber defense.
  • Development of CT-GMARL, a novel architecture for asynchronous multi-agent defense.
  • Empirical results show significant performance improvements over existing MARL baselines.
  • CT-GMARL effectively restores more compromised services and validates Zero-Shot transfer capabilities.
Read more
Alleviating Community Fear in Disasters via Multi-Agent Actor-Critic Reinforcement Learning
Yashodhan D. Hakke, Almuatazbellah M. Boker, Lamine Mili, Michael von Spakovsky, Hoda Eldardiry
Reinforcement Learning Optimization Theory
  • Introduction of a control-affine extension to the CPS model for disaster resilience.
  • Development of a three-player non-zero-sum differential game to optimize resource deployment.
  • Application of online actor-critic reinforcement learning to solve the game and derive optimal policies.
  • Empirical results show significant reductions in community fear during disasters.
Read more
Distilling Genomic Models for Efficient mRNA Representation Learning via Embedding Matching
Rasched Haidari, Sam Martin, Maxime Allard
Efficient ML
  • Embedding-based distillation is more stable than logit-based methods for genomic models.
  • The HelixNano-mRNA model achieves a 200-fold reduction in size while maintaining state-of-the-art performance.
  • Utilizing intermediate latent representations is an effective strategy for distilling knowledge in biological models.
Read more
Truncated Rectified Flow Policy for Reinforcement Learning with One-Step Sampling
Xubin Zhou, Yipeng Yang, Zhan Li
Reinforcement Learning Generative Models Robotics
  • TRFP integrates flow matching into MaxEnt RL, overcoming challenges of likelihood computation and backpropagation instability.
  • The framework employs a hybrid architecture for sampling, allowing for tractable optimization and stable training.
  • Flow straightening regularization enables high-fidelity one-step inference while minimizing surrogate divergence error.
  • Empirical results show TRFP achieves state-of-the-art performance on MuJoCo benchmarks and competes well with existing diffusion policies.
Read more
CORA: Conformal Risk-Controlled Agents for Safeguarded Mobile GUI Automation
Yushi Feng, Junye Du, Qifan Wang, Zizhan Ma, Qian Niu, Yutaka Matsuo, Long Feng, Lequan Yu
Computer Vision NLP Multimodal
  • CORA provides a formalized safety mechanism for mobile GUI automation using VLMs.
  • The framework employs a Guardian model for risk estimation and a Diagnostician for intervention recommendations.
  • CORA introduces a user-tunable execute/abstain threshold based on Conformal Risk Control.
  • The Phone-Harm benchmark is established to evaluate mobile safety violations in real-world settings.
Read more
Alloc-MoE: Budget-Aware Expert Activation Allocation for Efficient Mixture-of-Experts Inference
Baihui Liu, Kaiyuan Tian, Wei Wang, Zhaoning Zhang, Linbo Qiao, Dongsheng Li
Large Language Models Efficient ML Optimization
  • Introduces the concept of activation budget for expert activations in MoE models.
  • Presents Alloc-MoE, a unified framework optimizing expert allocation at both layer and token levels.
  • Alloc-L uses sensitivity profiling and dynamic programming for layer-level allocation.
  • Alloc-T dynamically redistributes expert activations based on routing scores.
Read more
Event-Centric World Modeling with Memory-Augmented Retrieval for Embodied Decision-Making
Fan Zhaowen
Robotics Interpretability Reinforcement Learning
  • Introduction of an event-centric framework for world modeling in autonomous agents.
  • Utilization of memory-augmented retrieval for decision-making based on prior experiences.
  • Integration of physics-informed knowledge to enhance maneuver selection consistency.
  • Demonstrated effectiveness in UAV flight scenarios under real-time control constraints.
Read more
GIRL: Generative Imagination Reinforcement Learning via Information-Theoretic Hallucination Control
Prakul Sunil Hiremath
Reinforcement Learning Generative Models Robotics
  • GIRL introduces a cross-modal grounding mechanism to maintain semantic consistency in imagined trajectories.
  • An uncertainty-adaptive trust-region bottleneck is used to control imagination drift based on real-environment feedback.
  • The framework shows a 38-61% reduction in latent rollout drift compared to DreamerV3.
  • GIRL achieves higher asymptotic returns with 40-55% fewer environment steps on long-horizon tasks.
Read more
Every Response Counts: Quantifying Uncertainty of LLM-based Multi-Agent Systems through Tensor Decomposition
Tiejin Chen, Huaiyuan Yao, Jia Chen, Evangelos E. Papalexakis, Hua Wei
NLP Large Language Models
  • Introduces the first systematic definition of uncertainty quantification for LLM-based multi-agent systems.
  • Presents MATU, a tensor decomposition-based framework for holistic uncertainty estimation.
  • Addresses unique sources of uncertainty in multi-agent systems, including tool usage and multi-step reasoning.
  • Demonstrates the effectiveness of MATU through extensive experiments across diverse tasks.
Read more
An Imperfect Verifier is Good Enough: Learning with Noisy Rewards
Andreas Plesner, Francisco Guzmán, Anish Athalye
Reinforcement Learning Large Language Models Theory
  • RLVR is robust to noise in verifier accuracy, with noise rates up to 15% showing minimal impact on performance.
  • The study emphasizes the importance of precision over recall in the context of verifier accuracy.
  • Imperfect verification does not pose a fundamental barrier to effective RLVR training.
  • Findings suggest that engineering efforts to improve verifier accuracy beyond a certain point yield diminishing returns.
Read more
Wireless Communication Enhanced Value Decomposition for Multi-Agent Reinforcement Learning
Diyi Hu, Bhaskar Krishnamachari
Reinforcement Learning Graph Learning Robotics
  • CLOVER leverages realistic wireless communication channels to enhance cooperation in MARL.
  • The GNN-based centralized mixer is conditioned on the communication graph, allowing for better credit assignment.
  • The framework achieves significant performance improvements over traditional methods like VDN and QMIX.
  • Agents learn effective communication strategies, adapting to varying channel conditions.
Read more
$p1$: Better Prompt Optimization with Fewer Prompts
Zhaolin Gao, Yu (Sid) Wang, Bo Liu, Thorsten Joachims, Kianté Brantley, Wen Sun
NLP Large Language Models Reinforcement Learning Optimization
  • Prompt optimization's effectiveness is influenced by the variance among system prompts versus responses.
  • Increasing the number of user prompts can reduce the effectiveness of prompt optimization.
  • The proposed p1 method filters user prompts to enhance the optimization signal.
  • p1 shows substantial improvements over traditional prompt optimization methods.
Read more
Shift- and stretch-invariant non-negative matrix factorization with an application to brain tissue delineation in emission tomography data
Anders S. Olsen, Miriam L. Navarro, Claus Svarer, Jesper L. Hinrich, Morten Mørup, Gitte M. Knudsen
Time Series
  • Introduction of a shift- and stretch-invariant NMF framework for dynamic neuroimaging data.
  • The model estimates both temporal shifts and stretching effects, improving the accuracy of TAC analysis.
  • Implementation in the frequency domain enhances computational efficiency.
  • Validation on synthetic and real SPECT data shows improved delineation of brain tissue structures.
Read more
Stochastic-Dimension Frozen Sampled Neural Network for High-Dimensional Gross-Pitaevskii Equations on Unbounded Domains
Zhangyong Liang
Optimization Efficient ML Theory
  • Introduces SD-FSNN, a neural network that is unbiased and dimension-independent.
  • Random sampling of weights and biases leads to faster training and improved accuracy.
  • Employs adaptive ODE solvers for effective temporal evolution and causality.
  • Integrates mechanisms for mass normalization and energy conservation.
Read more
Guardian-as-an-Advisor: Advancing Next-Generation Guardian Models for Trustworthy LLMs
Yue Huang, Haomin Zhuang, Jiayi Ye, Han Bao, Yanbo Wang, Hang Hua, Siyuan Wu, Pin-Yu Chen, Xiangliang Zhang
NLP Large Language Models Reinforcement Learning
  • Introduction of the Guardian-as-an-Advisor (GaaA) framework for LLMs.
  • Development of GuardSet, a large-scale dataset for training guardian models.
  • GuardAdvisor model trained to provide risk labels and explanations, enhancing response quality.
  • GaaA reduces unnecessary refusals while maintaining low latency.
Read more