AI-generated summaries

Today's ML research,
without the noise.

Daily summaries of the latest machine learning papers from arXiv, processed every 8 hours.

67 Papers today
8h Update frequency
7 Days of history
Rethinking Token-Level Credit Assignment in RLVR: A Polarity-Entropy Analysis
Yuhang He, Haodong Wu, Siyi Liu, Hongyu Ge, Hange Zhou, Keyi Wu, Zhuo Zheng, Qihong Lin, Zixin Zhong, Yongqi Zhang
Reinforcement Learning Large Language Models Optimization
  • Introduces the Four Quadrant Decomposition framework for analyzing token updates in RLVR.
  • Establishes a theoretical upper bound on token credit based on entropy using Conditional Mutual Information.
  • Demonstrates that reasoning improvements are primarily driven by high-entropy tokens.
  • Proposes Entropy-Aware Policy Optimization (EAPO) to optimize token-level learning signals.
Read more
Physics-Informed State Space Models for Reliable Solar Irradiance Forecasting in Off-Grid Systems
Mohammed Ezzaldin Babiker Abdullah
Time Series Efficient ML Theory
  • Introduction of the Physics-Informed State Space Model (PISSM) for solar irradiance forecasting.
  • Utilization of dynamic Hankel matrix embedding to filter noise from meteorological data.
  • Replacement of heavy RNNs and attention mechanisms with a Linear State Space Model for efficiency.
  • Implementation of a Physics-Informed Gating mechanism to ensure predictions adhere to physical laws.
Read more
Bringing Value Models Back: Generative Critics for Value Modeling in LLM Reinforcement Learning
Zikang Shan, Han Zhong, Liwei Wang, Li Zhao
Reinforcement Learning Large Language Models Generative Models
  • Identifies limitations of traditional discriminative critics in LLM reinforcement learning.
  • Introduces Generative Actor-Critic (GenAC) to enhance value modeling through chain-of-thought reasoning.
  • Implements In-Context Conditioning for better alignment between critic and actor during training.
  • Demonstrates superior performance in mathematical reasoning benchmarks compared to existing methods.
Read more
INCRT: An Incremental Transformer That Determines Its Own Architecture
Giansalvo Cirrincione
NLP Theory Efficient ML
  • INCRT dynamically adjusts its architecture during training, addressing structural redundancy in Transformers.
  • The model starts with a single attention head and adds or prunes heads based on real-time performance metrics.
  • Two theorems underpin the architecture's design, ensuring minimal and sufficient configurations.
  • Experimental validation shows INCRT can outperform BERT-base on specific tasks with fewer parameters.
Read more
Vestibular reservoir computing
Smita Deb, Shirin Panahi, Mulugeta Haile, Ying-Cheng Lai
Time Series Efficient ML Theory
  • Introduction of an uncoupled reservoir topology for reservoir computing.
  • Derivation of a memory capacity formula for linear reservoirs.
  • Demonstration of performance equivalence between uncoupled and fully coupled networks.
  • Exploration of the effects of reservoir size on predictive performance.
Read more
Are Independently Estimated View Uncertainties Comparable? Unified Routing for Trusted Multi-View Classification
Yilin Zhang, Cai Xu, Haishun Chen, Ziyu Guan, Wei Zhao
Multimodal
  • Identifies the fragility of the assumption that independently estimated view uncertainties are comparable.
  • Proposes TMUR, which decouples evidence extraction from fusion arbitration to improve multi-view classification.
  • Employs a unified router to generate sample-level expert weights based on global context.
  • Demonstrates through experiments that TMUR consistently outperforms existing methods in classification performance and reliability.
Read more
A Comparative Theoretical Analysis of Entropy Control Methods in Reinforcement Learning
Ming Lei, Christophe Baehr
Reinforcement Learning Large Language Models Theory
  • Traditional entropy regularization can lead to suboptimal policies due to persistent bias.
  • Covariance-based methods selectively regularize high-covariance tokens, achieving better performance.
  • The paper establishes a unified framework for understanding entropy dynamics in RL.
  • Covariance-based methods maintain stability margins, crucial for reasoning tasks.
Read more
Rethinking the Diffusion Model from a Langevin Perspective
Candi Zheng, Yuan Lan
Generative Models Theory Optimization
  • Introduces a Langevin perspective to simplify the understanding of diffusion models.
  • Unifies ODE-based and SDE-based diffusion models under a single framework.
  • Demonstrates the theoretical superiority of diffusion models compared to ordinary VAEs.
  • Clarifies the equivalence of flow matching, denoising, and score matching under maximum likelihood.
Read more
Silhouette Loss: Differentiable Global Structure Learning for Deep Representations
Matheus Vinícius Todescato, Joel Luís Carbonera
Computer Vision Optimization Theory
  • Introduction of Soft Silhouette Loss as a differentiable objective for representation learning.
  • The loss encourages intra-class compactness and inter-class separation without increasing computational complexity.
  • Combining Soft Silhouette Loss with cross-entropy and supervised contrastive learning yields superior performance.
  • Empirical results show consistent improvements across multiple image classification benchmarks.
Read more
WOODELF-HD: Efficient Background SHAP for High-Depth Decision Trees
Ron Wettenstein, Alexander Nadel, Udi Boker
Efficient ML Interpretability
  • WOODELF-HD improves the computational efficiency of Background SHAP for high-depth decision trees.
  • The algorithm reduces the preprocessing bottleneck from cubic to quadratic complexity with respect to tree depth.
  • It enables exact SHAP value computation for decision trees with depths up to 21, overcoming limitations of previous methods.
  • Significant speedups (up to 162×) are achieved over existing state-of-the-art algorithms for deep trees.
Read more
RTMC: Step-Level Credit Assignment via Rollout Trees
Tao Wang, Suhang Zheng, Xiaoxiao Xu
Reinforcement Learning Large Language Models Optimization
  • RTMC enables fine-grained credit assignment without a critic network.
  • The state-action signature system compresses interaction histories for efficient state matching.
  • Empirical results show a significant performance improvement over existing methods.
  • The approach addresses the limitations of traditional critic-free methods in multi-step RL.
Read more
EngageTriBoost: Predictive Modeling of User Engagement in Digital Mental Health Intervention Using Explainable Machine Learning
Ha Na Cho, Daniel Eisenberg, Cheryl King, Kai Zheng
Interpretability
  • EngageTriBoost (ETB) is an explainable ensemble ML framework for predicting user engagement in DMHI.
  • ETB achieved up to 84% accuracy in predicting message posting, outperforming individual models.
  • The study emphasizes interpretability and transparency in ML applications for mental health.
  • SHAP was used to identify key behavioral and demographic factors associated with user engagement.
Read more
Muon$^2$: Boosting Muon via Adaptive Second-Moment Preconditioning
Ziyue Liu, Ruijie Zhang, Zhengyang Wang, Yequan Zhao, Yupeng Su, Zi Yang, Zheng Zhang
Optimization Large Language Models Efficient ML
  • MUON2 improves the spectral properties of the momentum matrix, enhancing convergence speed.
  • The introduction of adaptive second-moment preconditioning leads to better optimization dynamics.
  • MUON2 and its factorized variant, MUON2-F, consistently outperform previous optimizers with reduced computational costs.
  • The method is validated through extensive experiments on large-scale models, showing significant efficiency gains.
Read more
Robust Adversarial Policy Optimization Under Dynamics Uncertainty
Mintae Kim, Koushil Sreenath
Reinforcement Learning Robotics Optimization
  • Introduces Robust Adversarial Policy Optimization (RAPO) to address dynamics uncertainty in RL.
  • Combines trajectory-level robustness through AdvNet with model-level robustness via Boltzmann reweighting.
  • Demonstrates improved resilience to uncertainty and generalization to out-of-distribution dynamics.
  • Maintains dual tractability while enhancing performance on in-distribution tasks.
Read more
Skill-SD: Skill-Conditioned Self-Distillation for Multi-turn LLM Agents
Hao Wang, Guozhi Wang, Han Xiao, Yufeng Zhou, Yue Pan, Jichao Wang, Ke Xu, Yafei Wen, Xiaohu Ruan, Xiaoxin Chen, Honggang Qi
Large Language Models Reinforcement Learning NLP
  • Skill-SD introduces dynamic, trajectory-derived natural language skills as a teacher signal for self-distillation.
  • An importance-weighted reverse-KL loss is developed to correct gradient biases during training.
  • Dynamic synchronization of the teacher model with the student is crucial for maintaining training stability.
  • Skill-SD outperforms traditional RL methods, showing significant improvements in task performance.
Read more
Temporal Dropout Risk in Learning Analytics: A Harmonized Survival Benchmark Across Dynamic and Early-Window Representations
Rafael da Silva, Jeff Eicher, Gregory Longo
Time Series
  • Introduces a harmonized survival benchmark for dropout risk modeling in Learning Analytics.
  • Demonstrates that temporal and behavioral signals are more predictive of dropout risk than static demographic factors.
  • Highlights the need for calibration and interpretability in predictive models beyond mere discrimination.
  • Finds that Random Survival Forest and Poisson Piecewise-Exponential are top performers in their respective model arms.
Read more
Act or Escalate? Evaluating Escalation Behavior in Automation with Language Models
Matthew DosSantos DiSorbo, Harang Ju
Large Language Models NLP Theory
  • Escalation behavior in LLMs is critical for effective automation and varies significantly across models.
  • Models exhibit miscalibration in their self-assessment of accuracy, affecting their decision-making.
  • Interventions such as supervised fine-tuning on chain-of-thought targets can improve escalation decisions.
  • The study highlights the need for careful characterization of model-specific escalation behavior prior to deployment.
Read more
On the Spectral Geometry of Cross-Modal Representations: A Functional Map Diagnostic for Multimodal Alignment
Krisanu Sarkar
Multimodal
  • First application of functional maps to multimodal neural representation alignment.
  • Evidence that independently pretrained vision and language encoders develop similar spectral complexity.
  • Identification of the spectral complexity–orientation gap in cross-modal representations.
  • Introduction of three new diagnostics for evaluating cross-modal representation compatibility.
Read more
NeuroFlow: Toward Unified Visual Encoding and Decoding from Neural Activity
Weijian Mai, Mu Nan, Yu Zhu, Jiahang Cao, Rui Zhang, Yuqin Dai, Chunfeng Song, Andrew F. Luo, Jiamin Wu
Multimodal
  • NeuroFlow is the first unified model for visual encoding and decoding from neural activity.
  • It incorporates NeuroVAE for structured latent space modeling and XFM for consistent flow learning.
  • The framework achieves superior performance and parameter efficiency compared to isolated methods.
  • NeuroFlow captures consistent neural activation patterns, enhancing understanding of visual perception.
Read more
Structural Gating and Effect-aligned Lag-resolved Temporal Causal Discovery Framework with Application to Heat-Pollution Extremes
Rui Chen, Jinsong Wu
Time Series Graph Learning Interpretability
  • Introduction of SGED-TCD framework for temporal causal discovery.
  • Application to heatwave and air pollution extremes in China reveals significant causal relationships.
  • Framework improves interpretability and robustness of causal graphs.
  • Demonstrates distinct regional and seasonal heterogeneity in causal pathways.
Read more
CARE-ECG: Causal Agent-based Reasoning for Explainable and Counterfactual ECG Interpretation
Elahe Khatibi, Ziyu Wang, Ankita Sharma, Krishnendu Chakrabarty, Sanaz Rahimi Moosavi, Farshad Firouzi, Amir Rahmani
Large Language Models Interpretability Time Series
  • CARE-ECG integrates causal reasoning into ECG interpretation, improving explainability and counterfactual analysis.
  • The framework encodes ECG signals into structured latent biomarkers, enhancing the interpretability of physiological factors.
  • CARE-ECG demonstrates significant improvements in diagnostic accuracy and reduces hallucinations in outputs.
  • The system supports rigorous evaluation through causal graph inference and counterfactual assessments.
Read more
Structured Exploration and Exploitation of Label Functions for Automated Data Annotation
Phong Lam, Ha-Linh Nguyen, Thu-Trang Nguyen, Son Nguyen, Hieu Dinh Vo
NLP Efficient ML Theory
  • Introduces EXPONA, a novel framework for automated data annotation using label functions.
  • Balances diversity and reliability in LF generation through a two-phase exploration and exploitation process.
  • Achieves near-complete label coverage and significantly improves weak label quality compared to existing methods.
  • Demonstrates substantial downstream performance gains across diverse classification tasks.
Read more
Exploring the impact of fairness-aware criteria in AutoML
Joana Simões, João Correia
Optimization
  • Integrating fairness metrics into AutoML can significantly improve fairness outcomes.
  • A trade-off exists between predictive performance and fairness, with a noted decrease in predictive power when fairness is prioritized.
  • The study employs a multi-criteria optimization approach to balance fairness and performance metrics.
  • Fairness-aware AutoML can lead to simpler and more efficient ML solutions.
Read more
Integrated electro-optic attention nonlinearities for transformers
Luis Mickeler, Kai Lion, Alfonso Nardi, Jost Kellner, Pierre Didier, Bhavin J. Shastri, Niao He, Rachel Grange
Computer Vision NLP Efficient ML
  • Proposes analog nonlinearities using TFLN MZMs to replace traditional Softmax in Transformers.
  • Demonstrates competitive accuracy in Vision Transformers and Large Language Models with 4-bit quantization.
  • Characterizes system performance under high encoding speeds and various noise conditions.
  • Addresses the computational bottleneck of Softmax operations in neural networks.
Read more
A Hybrid Intelligent Framework for Uncertainty-Aware Condition Monitoring of Industrial Systems
Maryam Ahang, Todd Charter, Masoud Jalayer, Homayoun Najjaran
Time Series
  • Development of a hybrid condition monitoring framework integrating data-driven and physics-based approaches.
  • Exploration of two hybrid integration strategies: feature-level fusion and model-level ensemble.
  • Demonstrated improvements in diagnostic accuracy and decision reliability through hybridization.
  • Application of conformal prediction for effective uncertainty quantification.
Read more
MoEITS: A Green AI approach for simplifying MoE-LLMs
Luis Balderas, Miguel Lastra, José M. Benítez
Large Language Models Efficient ML Theory
  • Introduction of MoEITS, a simplification algorithm for MoE-LLMs based on information theory.
  • Utilization of normalized mutual information to detect redundancy among experts.
  • Extensive empirical evaluation shows MoEITS outperforms existing pruning methods.
  • The method contributes to reducing computational burden and energy consumption in AI systems.
Read more
Virtual Smart Metering in District Heating Networks via Heterogeneous Spatial-Temporal Graph Neural Networks
Keivan Faghih Niresi, Christian Møller Jensen, Carsten Skovmose Kallesøe, Rafael Wisniewski, Olga Fink
Graph Learning Time Series Optimization
  • Introduction of HSTGNN for virtual smart metering in district heating networks.
  • Development of a controlled laboratory dataset for benchmarking virtual sensing methods.
  • Significant performance improvement over existing data-driven methods.
  • Joint modeling of cross-variable and spatial correlations in thermal and hydraulic states.
Read more
Consensus-based Recursive Multi-Output Gaussian Process
Yogesh Prasanna Kumar Rao, Tamas Keviczky, Raj Thilak Rajan
Robotics Efficient ML Theory
  • CRMGP combines recursive updates with distributed information fusion to enhance scalability.
  • The framework supports bounded per-step computation, making it suitable for real-time applications.
  • It preserves cross-output correlations, allowing for improved performance in multi-output tasks.
  • Experiments show that CRMGP outperforms traditional centralized Gaussian process models in predictive accuracy and uncertainty calibration.
Read more
Layerwise Dynamics for In-Context Classification in Transformers
Patrick Lutz, Themistoklis Haris, Arjun Chandra, Aditya Gangrade, Venkatesh Saligrama
Theory Interpretability Large Language Models
  • Enforcing symmetry in transformers enhances interpretability and reveals the underlying algorithmic structure.
  • A closed-form layerwise recursion is derived, demonstrating coupled dynamics between feature and label geometries.
  • The symmetry-enforced approach predicts behavior across various classification tasks, including semi-supervised learning.
  • The study provides a framework for understanding transformer dynamics beyond traditional optimization abstractions.
Read more
Offline Local Search for Online Stochastic Bandits
Gerdus Benadè, Rathish Das, Thomas Lavastida
Optimization Theory
  • Introduces a framework for converting offline local search algorithms into online stochastic bandit algorithms.
  • Achieves O(log3 T) regret, improving upon existing frameworks that yield polynomial regret.
  • Demonstrates flexibility by applying the framework to various combinatorial optimization problems.
  • Establishes conditions for local search neighborhoods to ensure effective online performance.
Read more
Task2vec Readiness: Diagnostics for Federated Learning from Pre-Training Embeddings
Cristiano Mafuz, Rodrigo Silva
Federated Learning
  • Introduction of Task2Vec Readiness as a diagnostic tool for federated learning.
  • Utilization of unsupervised metrics derived from Task2Vec embeddings to assess federation alignment.
  • Demonstrated strong correlation between readiness indices and final FL performance across various datasets.
  • Framework provides actionable guidance for client selection in heterogeneous federations.
Read more
ECHO: Efficient Chest X-ray Report Generation with One-step Block Diffusion
Lifeng Chen, Tianqi You, Hao Liu, Zhimin Bao, Jile Jiao, Xiao Han, Zhicai Ou, Tao Sun, Xiaofeng Mou, Xiaojie Jin, Yi Xu
Computer Vision NLP Generative Models
  • ECHO achieves efficient one-step report generation for chest X-rays, significantly reducing inference latency.
  • The Direct Conditional Distillation (DCD) framework enables coherent outputs by addressing mean-field bias.
  • Response-Asymmetric Diffusion (RAD) enhances training efficiency and model effectiveness.
  • ECHO surpasses existing autoregressive methods in performance metrics while maintaining clinical accuracy.
Read more
From Dispersion to Attraction: Spectral Dynamics of Hallucination Across Whisper Model Scales
Ivan Viakhirev, Kirill Borodin, Grach Mkrtchian
Audio & Speech Theory Interpretability
  • Introduction of the Spectral Sensitivity Theorem to explain hallucinations in ASR models.
  • Identification of two regimes: Structural Disintegration in smaller models and Compression-Seeking Attractor in larger models.
  • Validation of theoretical predictions through eigenspectral analysis of Whisper models under adversarial stress.
  • Demonstration that standard performance metrics may not adequately predict hallucination onset.
Read more
Toward World Models for Epidemiology
Zeeshan Memon, Yiqi Su, Christo Kurisummoottil Thomas, Walid Saad, Liang Zhao, Naren Ramakrishnan
Time Series
  • Introduces a conceptual framework for epidemiological world models.
  • Reframes epidemic decision-making to incorporate latent states and human behavior.
  • Presents three case studies demonstrating the utility of world models in policy analysis.
  • Highlights the limitations of traditional epidemiological models in dynamic environments.
Read more
Bringing Clustering to MLL: Weakly-Supervised Clustering for Partial Multi-Label Learning
Yu Chen, Weijun Lv, Yue Huang, Xuhuan Zhu, Fang Li
Theory Optimization
  • Introduction of a novel membership matrix decomposition that resolves incompatibility between clustering and multi-label scenarios.
  • Development of a three-stage weakly-supervised clustering framework that optimizes pseudo-labels and class prototypes.
  • Implementation of an adaptive confidence mechanism that adjusts supervision strength based on prototype-distance relationships.
  • Demonstration of superior performance over existing methods on multiple datasets.
Read more
Heterogeneous Connectivity in Sparse Networks: Fan-in Profiles, Gradient Hierarchy, and Topological Equilibria
Nikodem Tomczak
Theory Efficient ML Optimization
  • Introduction of Profiled Sparse Networks (PSN) for structured heterogeneous sparsity.
  • Static connectivity structures do not significantly affect accuracy at matched parameter counts.
  • Fan-in coefficient of variation (CV) predicts gradient concentration, indicating structural importance.
  • Lognormal initialization based on equilibrium fan-in distribution outperforms standard methods.
Read more
Below-ground Fungal Biodiversity Can be Monitored Using Self-Supervised Learning Satellite Features
Robin Young, Michael E. Van Nuland, E. Toby Kiers, Tomáš Větrovský, Petr Kohout, Petr Baldrian, Srinivasan Keshav
Multimodal Time Series Efficient ML
  • Self-supervised learning (SSL) can effectively predict below-ground ectomycorrhizal fungal richness using satellite imagery.
  • The proposed method achieves a 10,000-fold increase in spatial resolution compared to traditional biodiversity monitoring techniques.
  • SSL-derived features are more informative than conventional climate, soil, and land cover datasets for predicting fungal diversity.
  • The study enables temporal monitoring of fungal biodiversity, revealing trends in diversity loss in ancient forests.
Read more
QuanBench+: A Unified Multi-Framework Benchmark for LLM-Based Quantum Code Generation
Ali Slim, Haydar Hamieh, Jawad Kotaich, Yehya Ghosn, Mahdi Chehimi, Ammar Mohanna, Hasan Abed Al Kader Hammoud, Bernard Ghanem
Large Language Models
  • Introduction of QuanBench+, a multi-framework benchmark for quantum code generation.
  • Evaluation of LLMs across Qiskit, PennyLane, and Cirq with 42 aligned tasks.
  • Performance metrics include Pass@1, Pass@5, and feedback-based repair outcomes.
  • Results show that while models perform better with feedback, they still struggle with cross-framework reliability.
Read more
A Closer Look at the Application of Causal Inference in Graph Representation Learning
Hang Gao, Kunyu Li, Huang Hong, Baoquan Cui, Fengge Wu
Graph Learning Theory
  • Aggregation of graph elements into single causal variables violates causal inference assumptions.
  • A new theoretical model is proposed that adheres to the premises of causal inference.
  • A synthetic dataset mimicking real-world causal structures is created for empirical validation.
  • A plug-and-play causal modeling enhancement module is developed for graph learning pipelines.
Read more
Transformers Learn Latent Mixture Models In-Context via Mirror Descent
Francesco D'Angelo, Nicolas Flammarion
NLP Large Language Models Theory
  • Introduces a framework for in-context learning based on latent variables using Mixture of Transition Distributions.
  • Demonstrates that transformers can implement Mirror Descent to learn latent mixture weights from context.
  • Proves that the one-step estimator from the transformer is a first-order approximation of the Bayes-optimal predictor.
  • Empirical results show that transformers trained from scratch match predictive distributions and attention patterns consistent with the proposed framework.
Read more
Tracking High-order Evolutions via Cascading Low-rank Fitting
Zhao Song
Generative Models Theory Efficient ML
  • Introduces cascading low-rank fitting for modeling high-order dynamics in generative models.
  • Proves that under linear decomposability, the ranks of high-order derivatives are monotonically non-increasing.
  • Presents a computationally efficient algorithm for implementing the proposed method.
  • Demonstrates applicability to modern attention mechanisms in generative frameworks.
Read more
Structural Consequences of Policy-Based Interventions on the Global Supply Chain Network
Lea Karbevska, Liming Xu, Zehui Dai, Sara AlMahri, Alexandra Brintrup
Theory Optimization Graph Learning
  • Friendshoring increases globalization by enhancing supply links among allied countries.
  • Country Plus One policy enhances network density through redundant links.
  • Reshoring creates challenges in the EV sector due to irreplaceable products.
  • The impact of these policies varies across different industries.
Read more
Integrating SAINT with Tree-Based Models: A Case Study in Employee Attrition Prediction
Adil Derrazi, Javad Pourmostafa Roshan Sharami
Interpretability
  • Tree-based models (XGBoost and LightGBM) outperform standalone SAINT and hybrid models in predictive accuracy.
  • Hybrid models did not improve performance and sometimes performed worse than standalone SAINT.
  • Tree-based models demonstrated strong generalization, while hybrid models showed performance degradation.
  • SAINT embeddings may not align well with tree-based classifiers optimized for structured data.
Read more
Joint Interference Detection and Identification via Adversarial Multi-task Learning
H. Xu, B. He, S. Wang
Theory
  • Introduces a theoretically grounded MTL framework for joint interference detection and identification.
  • Derives an upper bound for weighted expected loss linked to task similarity using Wasserstein distance.
  • Develops AMTIDIN, which utilizes adversarial training to enhance task correlation modeling.
  • Quantitative analysis reveals significant feature overlap between modulation and interference identification tasks.
Read more
Mycelium-Index: A Streaming Approximate Nearest Neighbor Index with Myelial Edge Decay, Traffic-Driven Reinforcement, and Adaptive Living Hierarchy
Anton Pakhunov
Graph Learning Efficient ML Optimization
  • Mycelium-Index adapts its structure dynamically based on query traffic, improving efficiency and memory usage.
  • The system achieves high recall rates while significantly reducing RAM usage compared to existing methods.
  • A hybrid deletion strategy enhances performance by efficiently managing cold and hub nodes.
  • The study reveals that topological mechanisms are more effective than geometric ones for high-dimensional ANN repair.
Read more
Explainable Human Activity Recognition: A Unified Review of Concepts and Mechanisms
Mainak Kundu, Catherine Chen, Rifatul Islam, Ismail Uysal, Ria Kanjilal
Interpretability Time Series Multimodal
  • Introduces a unified framework for understanding explainability in HAR systems.
  • Presents a mechanism-centric taxonomy categorizing XAI-HAR methods.
  • Addresses the complexities of temporal, multimodal, and semantic aspects in HAR.
  • Identifies key challenges in the deployment of reliable XAI-HAR systems.
Read more
Hierarchical Flow Decomposition for Turning Movement Prediction at Signalized Intersections
Md Atiqur Rahman Mallick, Kamrul Hasan, Pulock Das, Liang Hong, S M Shazzad Rassel
Time Series
  • Introduction of HFD-TM, a hierarchical framework for turning movement prediction.
  • Utilization of corridor through-flows to improve prediction accuracy for turning movements.
  • Implementation of a physics-informed loss function to enforce flow conservation.
  • Demonstrated significant performance improvements over existing models.
Read more
SCOPE: Signal-Calibrated On-Policy Distillation Enhancement with Dual-Path Adaptive Weighting
Binbin Zheng, Xing Ma, Yiheng Liang, Jingqing Ruan, Xiaoliang Fu, Kepeng Lin, Benchang Zhu, Ke Zeng, Xunliang Cai
Reinforcement Learning Large Language Models NLP
  • SCOPE introduces a dual-path adaptive framework for on-policy reinforcement learning.
  • The framework distinguishes between correct and incorrect trajectories to apply tailored supervision.
  • Empirical analysis reveals the importance of signal quality in OPD, leading to improved learning outcomes.
  • SCOPE achieves an average relative improvement of 11.42% in Avg@32 and 7.30% in Pass@32 over competitive baselines.
Read more
Preventing Latent Rehearsal Decay in Online Continual SSL with SOLAR
Giacomo Cignoni, Simone Magistri, Andrew D. Bagdanov, Antonio Carta
Computer Vision
  • Introduces the Latent Rehearsal Decay hypothesis to explain performance drops in OCSSL.
  • Develops two novel metrics, Overlap and Deviation, to diagnose latent space degradation.
  • Proposes SOLAR, a method that combines a Deviation-Aware Buffer and Overlap Loss for adaptive plasticity management.
  • Demonstrates SOLAR's effectiveness through extensive experiments, achieving state-of-the-art results.
Read more
Efficient Hierarchical Implicit Flow Q-learning for Offline Goal-conditioned Reinforcement Learning
Zhiqiang Dong, Teng Pang, Rongjian Xu, Guoqiang Wu
Reinforcement Learning Generative Models Robotics
  • Introduction of Hierarchical Implicit Flow Q-Learning (HIFQL) for offline GCRL.
  • Utilization of mean flow policies to enhance expressiveness and efficiency in hierarchical policy learning.
  • Implementation of a LeJEPA loss to improve goal representation and generalization.
  • Strong performance demonstrated on OGBench benchmark across various tasks.
Read more
Ranked Activation Shift for Post-Hoc Out-of-Distribution Detection
Gianluca Guglielmo, Marc Masana
Computer Vision
  • Introduces RAS, a hyperparameter-free method for OoD detection that enhances activation shifts.
  • Demonstrates consistent performance across different datasets and model architectures.
  • Identifies the limitations of existing scaling-based methods in handling unrectified activations.
  • Shows that both inhibitory and excitatory shifts independently improve OoD detection.
Read more
ANTIC: Adaptive Neural Temporal In-situ Compressor
Sandeep S. Cranganore, Andrei Bodnar, Gianluca Galleti, Fabian Paischer, Johannes Brandstetter
Efficient ML Time Series Theory
  • Introduction of ANTIC, a novel in-situ compression framework for multi-rate/stiff PDE simulations.
  • Utilization of physics-aware metrics for selecting salient temporal snapshots.
  • Implementation of neural spatial compression through continual fine-tuning of residuals.
  • Demonstrated significant storage reductions while preserving physics accuracy.
Read more
A Diffusion-Contrastive Graph Neural Network with Virtual Nodes for Wind Nowcasting in Unobserved Regions
Jie Shi, Siamak Mehrkanoon
Graph Learning Time Series
  • Introduces a novel framework (ContraVirt) for wind nowcasting in unobserved regions using virtual nodes.
  • Achieves a significant reduction in mean absolute error (MAE) of wind predictions by 30% to 46% compared to traditional methods.
  • Utilizes contrastive learning strategies to enhance model robustness and representation in data-scarce areas.
  • Grounded in geographic principles, the model effectively learns from neighboring observed regions.
Read more
A Temporally Augmented Graph Attention Network for Affordance Classification
Ami Chopra, Supriya Bordoloi, Shyamanta M. Hazarika
Graph Learning Time Series
  • Introduction of EEG-tGAT, a temporally enhanced GAT for EEG affordance classification.
  • Incorporation of temporal attention and dropout to address non-uniform temporal dynamics in EEG data.
  • Demonstrated improved classification performance over traditional GATv2 models.
  • Findings suggest that temporal modeling aligns better with the neurocognitive processes involved in affordance perception.
Read more
FluidFlow: a flow-matching generative model for fluid dynamics surrogates on unstructured meshes
David Ramos, Lucas Lacasa, Fermín Gutiérrez, Eusebio Valero, Gonzalo Rubio
Generative Models
  • FluidFlow utilizes conditional flow-matching for scalable fluid dynamics surrogate modeling.
  • The model operates directly on unstructured meshes without requiring mesh interpolation.
  • FluidFlow outperforms traditional multilayer perceptron models in accuracy and generalization.
  • The transformer architecture allows for efficient learning from large datasets.
Read more
Exact Certification of Neural Networks and Partition Aggregation Ensembles against Label Poisoning
Ajinkya Mohgaonkar, Lukas Gosch, Mahalakshmi Sabanayagam, Debarghya Ghoshdastidar, Stephan Günnemann
Theory Efficient ML
  • Introduces EnsembleCert, the first white-box certification framework for partition-aggregation ensembles.
  • Develops ScaLabelCert, enabling exact certification of neural networks against label-flipping attacks.
  • Demonstrates significant improvements in certified robustness over existing black-box methods.
  • Reduces the number of required partitions for effective certification, enhancing efficiency.
Read more
SLM Finetuning for Natural Language to Domain Specific Code Generation in Production
Renjini R. Nair, Damian K. Kowalczyk, Marco Gaudesi, Chhaya Methani
NLP Large Language Models Efficient ML
  • Fine-tuning SLMs significantly improves their performance for domain-specific code generation.
  • SLMs provide a resource-efficient alternative to LLMs, particularly in production environments with strict latency requirements.
  • The study demonstrates successful adaptation of fine-tuned models for customer-specific scenarios without performance loss.
  • Load testing and real-world deployment confirm the effectiveness of the proposed approach.
Read more
AdaCubic: An Adaptive Cubic Regularization Optimizer for Deep Learning
Ioannis Tsingalis, Constantine Kotropoulos, Corentin Briat
Optimization
  • AdaCubic adapts the cubic term weight dynamically, enhancing optimization efficiency.
  • Utilizes Hutchinson's method for Hessian approximation, reducing computational overhead.
  • Demonstrates superior performance compared to existing optimizers in multiple domains.
  • Does not require hyperparameter fine-tuning, making it user-friendly.
Read more
A Mechanistic Analysis of Looped Reasoning Language Models
Hugh Blayney, Álvaro Arroyo, Johan Obando-Ceron, Pablo Samuel Castro, Aaron Courville, Michael M. Bronstein, Xiaowen Dong
Large Language Models Theory Interpretability
  • Looped language models tend toward cyclic fixed-point behavior, stabilizing attention patterns.
  • Recurrent blocks learn stages of inference that closely resemble those in feedforward models.
  • Architectural choices significantly influence the emergence and stability of cyclic fixed points.
  • Empirical evidence shows that stable models maintain consistent inference stages, while unstable models deviate.
Read more
Gradient-Variation Regret Bounds for Unconstrained Online Learning
Yuheng Zhao, Andrew Jacobsen, Nicolò Cesa-Bianchi, Peng Zhao
Theory Optimization
  • Development of the first fully-adaptive algorithm for gradient-variation online learning in unbounded domains.
  • Introduction of a new definition of gradient variation that is effective for arbitrary comparators.
  • Algorithms achieve regret bounds that do not require prior knowledge of comparator norms or other parameters.
  • Efficient computation with closed-form updates, ensuring linear time complexity per round.
Read more
From Selection to Scheduling: Federated Geometry-Aware Correction Makes Exemplar Replay Work Better under Continual Dynamic Heterogeneity
Zhuang Qi, Ying-Peng Tang, Lei Meng, Guoqing Chao, Lei Wu, Han Yu, Xiangxu Meng
Federated Learning
  • FEAT addresses inter-client heterogeneity and class imbalance in exemplar replay-based FCL.
  • The method includes a geometric structure alignment for consistent feature representation across clients.
  • An energy-based correction improves model sensitivity to minority classes.
  • FEAT shows significant performance improvements over existing state-of-the-art methods.
Read more
Uncertainty-Aware Transformers: Conformal Prediction for Language Models
Abhiram Vellore, Niraj K. Jha
NLP Large Language Models Interpretability
  • Introduction of CONFIDE, a conformal prediction framework for transformer models.
  • Achieves up to 4.09% improvement in test accuracy and greater correct efficiency over existing methods.
  • Demonstrates that early and intermediate transformer layers provide better-calibrated representations.
  • Offers robustness and interpretability in high-stakes applications where traditional softmax-based uncertainty fails.
Read more
Fairboard: a quantitative framework for equity assessment of healthcare models
James K. Ruffle, Samia Mohinta, Chris Foulon, Mohamad Zeina, Zicheng Wang, Sebastian Brandner, Harpreet Hyare, Parashkev Nachev
Computer Vision
  • Fairboard provides a comprehensive framework for assessing equity in healthcare AI models.
  • Patient identity and clinical factors significantly influence model performance more than model architecture.
  • Spatial biases in model performance are identified, revealing compartment-specific inequities.
  • Newer models show improvements in equity, but none offer formal fairness guarantees.
Read more
Active Bayesian Inference for Robust Control under Sensor False Data Injection Attacks
Axel Andersson, György Dán
Robotics Graph Learning Theory
  • Introduces a bipartite graph model for perception pipelines in CPSs.
  • Develops the LASE-AD algorithm for maintaining beliefs over sensor attack states.
  • Proposes an active probing strategy to enhance detection of compromised sensors.
  • Demonstrates significant performance improvements over traditional methods in experimental settings.
Read more
Beyond Fixed False Discovery Rates: Post-Hoc Conformal Selection with E-Variables
Meiyi Zhu, Osvaldo Simeone
Theory
  • PH-CS allows for post-hoc adjustments to candidate selection based on observed data, overcoming the limitations of fixed FDR levels.
  • The method provides a path of candidate sets with associated FDP estimates, enabling flexible decision-making based on user-defined utility.
  • PH-CS guarantees that the average estimated FDP is a valid upper bound on the true FDR, ensuring statistical validity.
  • Experiments show that PH-CS can satisfy user utility constraints while maintaining competitive performance compared to traditional CS.
Read more
Transformers Learn the Optimal DDPM Denoiser for Multi-Token GMMs
Hongkang Li, Hancheng Min, Rene Vidal
Generative Models Theory Optimization
  • First convergence analysis for transformer-based diffusion models under DDPM loss.
  • Quantifies the number of tokens and training iterations needed for convergence.
  • Demonstrates that transformers can learn the oracle MMSE estimator for denoising.
  • Identifies the impact of data distribution characteristics on convergence.
Read more
PokeRL: Reinforcement Learning for Pokemon Red
Dheeraj Mudireddy, Sai Patibandla
Reinforcement Learning
  • PokeRL addresses the challenges of sparse rewards and partial observability in Pokemon Red.
  • The system incorporates mechanisms to prevent common pitfalls such as action loops and button spamming.
  • Training is structured as a curriculum over three specific early-game tasks.
  • The environment is designed to enhance agent robustness and transparency in learning.
Read more