AI-generated summaries

Today's ML research,
without the noise.

Daily summaries of the latest machine learning papers from arXiv, processed every 8 hours.

38 Papers today
8h Update frequency
7 Days of history
Dilated CNNs for Periodic Signal Processing: A Low-Complexity Approach
Eli Gildish, Michael Grebshtein, Igor Makienko
Time Series Efficient ML Audio & Speech
  • R-DCNN offers a low-complexity solution for denoising periodic signals.
  • The method requires only a single observation for training, enabling efficient generalization to other signals.
  • R-DCNN achieves performance comparable to classical autoregressive methods and conventional DCNNs.
  • The approach is particularly suited for resource-constrained environments like IoT devices.
Read more
Forget, Then Recall: Learnable Compression and Selective Unfolding via Gist Sparse Attention
Yuzhen Mao, Michael Y. Li, Emily B. Fox
NLP Large Language Models Efficient ML
  • Introduces a novel method for long-context modeling in LLMs using gist compression tokens.
  • Proposes selective unfolding via Gist Sparse Attention (GSA) to enhance attention efficiency.
  • Demonstrates significant performance improvements over existing compression and sparse attention methods.
  • Enables multi-resolution context access with logarithmic complexity through recursive gist construction.
Read more
Sink-Token-Aware Pruning for Fine-Grained Video Understanding in Efficient Video LLMs
Kibum Kim, Jiwan Kim, Kyle Min, Yueqi Wang, Jinyoung Moon, Julian McAuley, Chanyoung Park
Computer Vision Large Language Models Efficient ML
  • Introduction of sink tokens as a critical challenge in fine-grained video understanding.
  • Development of Sink-Token-aware Pruning (SToP) to effectively target and suppress sink tokens.
  • SToP significantly improves performance on fine-grained tasks while allowing for substantial token pruning.
  • Validation of SToP across diverse benchmarks, including hallucination evaluation and open-ended generation.
Read more
Decoupled Travel Planning with Behavior Forest
Duanyang Yuan, Sihang Zhou, Yanning Hou, Xiaoshu Chen, Haoyuan Chen, Ke Liang, Jiyuan Liu, Chuan Ma, Xinwang Liu, Jian Huang
NLP Large Language Models Optimization
  • Introduces the Behavior Forest method to decouple travel planning tasks.
  • Structures decision-making into parallel behavior trees for modular planning.
  • Integrates LLMs for localized reasoning within behavior tree nodes.
  • Demonstrates significant performance improvements over existing methods.
Read more
Clinically Interpretable Sepsis Early Warning via LLM-Guided Simulation of Temporal Physiological Dynamics
Weizhi Nie, Zhen Qu, Weijie Wang, Chunpei Li, Ke Lu, Bingyang Zhou, Hongzhi Yu
Large Language Models Time Series Interpretability
  • Introduces a novel LLM-guided framework for sepsis early warning that enhances clinical interpretability.
  • Combines spatiotemporal feature extraction with medical prompt engineering to improve prediction accuracy.
  • Achieves superior AUC scores compared to traditional models, demonstrating effectiveness in pre-onset prediction tasks.
  • Provides interpretable physiological trajectories that support clinical decision-making.
Read more
Low-Rank Adaptation Redux for Large Models
Bingcong Li, Yilang Zhang, Georgios B. Giannakis
Large Language Models Optimization Efficient ML
  • LoRA is a leading method for parameter-efficient fine-tuning of large models, significantly reducing computational and memory costs.
  • The paper categorizes advancements in LoRA into architectural design, optimization techniques, and applications across the model lifecycle.
  • Signal processing principles can enhance the understanding and development of LoRA methods, offering a structured approach to fine-tuning.
  • Emerging applications of LoRA extend beyond fine-tuning to include pre-training and deployment strategies.
Read more
Transferable SCF-Acceleration through Solver-Aligned Initialization Learning
Eike S. Eberhard, Viktor Kotsev, Timm Güthle, Stephan Günnemann
Optimization Efficient ML Theory
  • SAIL addresses the supervision problem in ML models for SCF initialization, improving convergence for larger molecules.
  • The introduction of the Effective Relative Iteration Count (ERIC) provides a more accurate measure of convergence efficiency.
  • SAIL achieves significant reductions in ERIC, outperforming previous state-of-the-art methods.
  • The method is applicable to both Hamiltonian and density matrix models, enhancing their performance in practical scenarios.
Read more
CAP: Controllable Alignment Prompting for Unlearning in LLMs
Zhaokun Wang, Jinyu Guo, Jingwen Pu, Hongli Pu, Meng Yang, Xunlei Chen, Jie Ou, Wenyi Li, Guangchun Luo, Wenhong Tian
Large Language Models Reinforcement Learning NLP
  • CAP is the first end-to-end trained prompt-driven unlearning framework for LLMs.
  • The framework utilizes reinforcement learning to optimize prompts for targeted knowledge suppression.
  • CAP achieves precise unlearning without the need for model parameter updates.
  • Extensive experiments show CAP outperforms existing methods in forgetting rate and retention accuracy.
Read more
Differentially Private Model Merging
Qichuan Yin, Manzil Zaheer, Tian Li
Theory Efficient ML Federated Learning
  • Introduces two data-independent algorithms for merging private models: random selection and linear combination.
  • Provides tailored privacy accounting using R´enyi differential privacy and privacy loss distributions.
  • Demonstrates the theoretical superiority of linear combination over random selection in terms of privacy/utility trade-off.
  • Validates the proposed methods through empirical evaluations on various datasets.
Read more
A Deep U-Net Framework for Flood Hazard Mapping Using Hydraulic Simulations of the Wupper Catchment
Christian Lammers, Fernando Arévalo, Leonie Märker-Neuhaus, Daniel Heinenberg, Christian Förster, Karl-Heinz Spies
Efficient ML
  • Development of a deep learning surrogate model for flood prediction using U-Net architecture.
  • The model provides a computationally efficient alternative to traditional hydraulic simulations.
  • Testing was conducted using hydraulic simulations from the Wupper catchment, yielding comparable results.
  • The framework aims to be generalizable across various topographies for broader application.
Read more
Interpretable Quantile Regression by Optimal Decision Trees
Valentin Lemaire, Gaël Aglin, Siegfried Nijssen
Interpretability
  • Introduces Quantile DL8.5 (QDL8.5) for optimal quantile regression trees.
  • Provides predictions for the complete conditional distribution of a target variable.
  • Enhances interpretability and robustness by learning multiple trees for different quantiles.
  • Achieves high accuracy with minimal computational overhead compared to single tree learning.
Read more
Geometric Monomial (GEM): a family of rational 2N-differentiable activation functions
Eylon E. Krause
Optimization Theory Efficient ML
  • Introduction of GEM, a family of rational 2N-differentiable activation functions.
  • Three variants of GEM are proposed: GEM, E-GEM, and SE-GEM, each addressing different optimization challenges.
  • N-ablation study reveals N=1 is optimal for CNNs while N=2 is better for transformers.
  • GEM outperforms GELU in specific benchmarks, achieving lower deficits and better performance metrics.
Read more
Validating a Deep Learning Algorithm to Identify Patients with Glaucoma using Systemic Electronic Health Records
John Xiang, Rohith Ravindranath, Sophia Y. Wang
Efficient ML
  • The GRA model effectively identifies high-risk glaucoma patients using systemic EHR data.
  • The model achieved an AUROC of 0.883 and a PPV of 0.657, indicating strong predictive performance.
  • Calibration of the model aligns with clinical risk assessments, enhancing its practical utility.
  • Fine-tuning the model on local data improved its performance, demonstrating the importance of external validation.
Read more
A temporal deep learning framework for calibration of low-cost air quality sensors
Arindam Sengupta, Tony Bush, Ben Marner, Jose Miguel Pérez, Soledad Le Clainche
Time Series
  • Introduces a deep learning framework for calibrating low-cost air quality sensors using LSTM networks.
  • Captures temporal dependencies and environmental effects, improving calibration accuracy over traditional methods.
  • Achieves regulatory compliance with significant reductions in uncertainty for calibrated pollutant measurements.
  • Utilizes advanced feature engineering to enhance model generalization across different temporal contexts.
Read more
Conditional anomaly detection with soft harmonic functions
Michal Valko, Branislav Kveton, Hamed Valizadegan, Gregory F. Cooper, Milos Hauskrecht
Graph Learning
  • Introduction of a non-parametric method for conditional anomaly detection using soft harmonic functions.
  • Regularization techniques to avoid misclassification of isolated and fringe points.
  • Development of a compact computation method for building a backbone graph to facilitate label propagation.
  • Demonstration of the method's efficacy on synthetic, UCI, and real-world datasets.
Read more
The Sample Complexity of Multicalibration
Natalie Collina, Jiuyao Lu, Georgy Noarov, Aaron Roth
Theory
  • Establishes the sample complexity of multicalibration as eΘ(ε−3) for certain group sizes.
  • Differentiates multicalibration from marginal calibration, which has a lower sample complexity of eΘ(ε−2).
  • Demonstrates that mean-ECE multicalibration is equally difficult in both batch and online settings.
  • Identifies a sharp threshold phenomenon in sample complexity when κ = 0.
Read more
The Recurrent Transformer: Greater Effective Depth and Efficient Decoding
Costin-Andrei Oncescu, Depen Morwani, Samy Jelassi, Alexandru Meterez, Mujin Kwun, Sham Kakade
NLP Large Language Models Efficient ML
  • Introduces the Recurrent Transformer architecture, enhancing effective depth and efficiency.
  • Emulates both conventional Transformer behavior and token-to-token recurrent updates.
  • Presents a tiling algorithm that reduces memory traffic and increases arithmetic intensity.
  • Demonstrates improved performance in cross-entropy loss with fewer layers compared to traditional Transformers.
Read more
Understanding and Mitigating Spurious Signal Amplification in Test-Time Reinforcement Learning for Math Reasoning
Yongcan Yu, Lingxiao He, Jian Liang, Kuangpu Guo, Meng Wang, Qianlong Xie, Xingxing Wang, Ran He
Reinforcement Learning Large Language Models NLP
  • Medium-frequency samples are identified as a major source of spurious reward signals.
  • Group-relative advantage normalization amplifies these spurious signals during optimization.
  • The DDRL framework introduces a balanced sampling strategy and debiased advantage estimation.
  • Extensive experiments show significant performance improvements over existing TTRL methods.
Read more
Sub-Token Routing in LoRA for Adaptation and Query-Aware KV Compression
Wei Jiang, Wei Wang
NLP Large Language Models Efficient ML
  • Introduces sub-token routing for finer control in transformer efficiency.
  • Presents a query-independent design that enhances language modeling quality.
  • Develops a query-aware design that maintains downstream performance under reduced KV budgets.
  • Demonstrates the complementary nature of token-level and sub-token-level routing.
Read more
Droplet-LNO: Physics-Informed Laplace Neural Operators for Accurate Prediction of Droplet Spreading Dynamics on Complex Surfaces
Ganesh Sahadeo Meshram, Partha Pratim Chakrabarti, Suman Chakraborty
Theory Efficient ML Optimization
  • Introduction of PI-LNO, a novel neural network architecture for droplet dynamics prediction.
  • Significant reduction in computation time, achieving a ∼23,400× speedup over traditional CFD methods.
  • Demonstrated superior performance with a mean R2 score of 0.9009 compared to existing models.
  • Incorporation of physics-informed constraints enhances model accuracy and physical interpretability.
Read more
JEPAMatch: Geometric Representation Shaping for Semi-Supervised Learning
Ali Aghababaei-Harandi, Aude Sportisse, Massih-Reza Amini
Computer Vision Theory Efficient ML
  • JEPAMatch addresses class imbalance and convergence issues in semi-supervised learning.
  • The method combines pseudo-labeling with geometric representation shaping in latent space.
  • Extensive experiments show JEPAMatch outperforms existing methods on multiple datasets.
  • The approach significantly accelerates convergence and reduces computational costs.
Read more
Synthetic Data in Education: Empirical Insights from Traditional Resampling and Deep Generative Models
Tapiwa Amion Chinodakufa, Ashfaq Ali Shafin, Khandaker Mamun Ahmed
Generative Models
  • Synthetic data generation can mitigate data scarcity and privacy issues in education.
  • Traditional resampling methods provide high utility but lack privacy protection.
  • Deep learning models offer better privacy guarantees but at a significant utility cost.
  • Variational Autoencoders are identified as the optimal balance between utility and privacy.
Read more
Dynamical Priors as a Training Objective in Reinforcement Learning
Sukesh Subaharan
Reinforcement Learning
  • DP-RL framework introduces an auxiliary loss to impose temporal structure in RL training.
  • The approach does not modify the reward structure, environment, or policy architecture.
  • Dynamical priors significantly alter decision trajectories, promoting temporally coherent behavior.
  • The study demonstrates that training objectives can control the temporal geometry of decision-making.
Read more
Supervised Learning Has a Necessary Geometric Blind Spot: Theory, Consequences, and Minimal Repair
Vishal Rajput
Theory
  • Theorem establishes a necessary geometric flaw in representations learned via ERM, termed the geometric blind spot.
  • Introduces the Trajectory Deviation Index (TDI) to measure geometric distortion, revealing limitations of existing metrics.
  • Confirms that the blind spot worsens with model scale and is amplified by task-specific fine-tuning.
  • Proposes a minimal fix (PMH) that effectively reduces the blind spot while maintaining performance.
Read more
Spectral Embeddings Leak Graph Topology: Theory, Benchmark, and Adaptive Reconstruction
Thinh Nguyen-Cong, Truong-Son Hy, Thang N. Dinh
Graph Learning Federated Learning Theory
  • Introduction of LoGraB, a benchmark for fragmented graph learning.
  • Development of AFR, an adaptive method for reconstructing noisy spectral fragments.
  • Establishment of the Spectral Leakage Proposition for polynomial-time graph recovery.
  • Demonstration of AFR's effectiveness in maintaining performance under privacy constraints.
Read more
Evaluating Post-hoc Explanations of the Transformer-based Genome Language Model DNABERT-2
Isabel Kurth, Paulo Yanez Sarmiento, Bernhard Y. Renard
Interpretability
  • Adaptation of AttnLRP for explaining DNABERT-2 model predictions.
  • Comparison of explanations from DNABERT-2 and a baseline CNN.
  • Demonstration that Transformer-based models can yield biologically relevant insights.
  • Evaluation of explanations using multiple metrics including sparsity and faithfulness.
Read more
Toward Efficient Membership Inference Attacks against Federated Large Language Models: A Projection Residual Approach
Guilin Deng, Silong Chen, Yuchuan Luo, Yi Liu, Songlei Wang, Zhiping Cai, Lin Liu, Xiaohua Jia, Shaojing Fu
Large Language Models Federated Learning NLP
  • ProjRes is the first projection residuals-based passive MIA specifically designed for FedLLMs.
  • The method achieves near 100% accuracy in membership inference, outperforming previous techniques.
  • ProjRes operates without the need for shadow models or auxiliary classifiers, enhancing efficiency.
  • The study reveals significant privacy vulnerabilities in FedLLMs, necessitating a reevaluation of their security measures.
Read more
Measure Twice, Click Once: Co-evolving Proposer and Visual Critic via Reinforcement Learning for GUI Grounding
Wenkai Wang, Xiyun Li, Hongcan Guo, Wenhao Yu, Tianqing Fang, Haitao Mi, Dong Yu, Shengyu Zhang
Reinforcement Learning Computer Vision Multimodal
  • Introduction of a Propose-then-Critic framework for GUI grounding.
  • Utilization of a co-evolutionary reinforcement learning strategy to enhance model capabilities.
  • Dynamic maturity-aware mechanism to balance prediction accuracy and diversity.
  • Significant improvements in grounding accuracy and critic reliability.
Read more
Generalizing Numerical Reasoning in Table Data through Operation Sketches and Self-Supervised Learning
Hanjun Cho, Gahyun Yoo, Hanseong Kim, Jay-Yoon Lee
NLP Large Language Models Efficient ML
  • Identifies three failure modes in numerical reasoning: reasoning inefficiency, data scarcity for logical supervision, and header dependency.
  • Introduces operation sketches to focus models on contextual reasoning rather than surface-level patterns.
  • Combines operation sketches with header anonymization and self-supervised learning in the TaNOS framework.
  • Achieves superior performance on FinQA with significantly less training data compared to traditional methods.
Read more
GFlowState: Visualizing the Training of Generative Flow Networks Beyond the Reward
Florian Holeczek, Andreas Hinterreiter, Alex Hernandez-Garcia, Marc Streit, Christina Humer
Generative Models Interpretability
  • GFlowState enhances the interpretability of GFlowNets by visualizing training dynamics.
  • The system provides multiple interactive views for analyzing sampling behavior and policy evolution.
  • Case studies show GFlowState's effectiveness in debugging and assessing GFlowNets.
  • The tool addresses the gap in existing visualization methods for GFlowNet training.
Read more
Transferable Physics-Informed Representations via Closed-Form Head Adaptation
Jian Cheng Wong, Isaac Yin Chung Lai, Pao-Hsiung Chiu, Chin Chun Ooi, Abhishek Gupta, Yew-Soon Ong
Theory Optimization Efficient ML
  • Introduction of Pi-PINN framework for transferable physics-informed representations.
  • Closed-form head adaptation significantly reduces computational costs for adapting to new PDE instances.
  • Improved generalization across PDE families through shared embedding learning.
  • Empirical results show substantial speed and accuracy improvements over traditional PINNs.
Read more
mcdok at SemEval-2026 Task 13: Finetuning LLMs for Detection of Machine-Generated Code
Adam Skurla, Dominik Macko, Jakub Simko
Large Language Models NLP Generative Models
  • The mcdok system is designed for multi-domain detection of machine-generated code.
  • It adapts the existing mdok approach for better code understanding.
  • The system is evaluated across three subtasks: binary detection, authorship detection, and hybrid code detection.
  • Results show competitive performance, but significant room for improvement remains.
Read more
A Scale-Adaptive Framework for Joint Spatiotemporal Super-Resolution with Diffusion Models
Max Defez, Filippo Quarenghi, Mathieu Vrac, Stephan Mandt, Tom Beucler
Generative Models Time Series Computer Vision
  • Introduces a scale-adaptive framework for joint spatiotemporal super-resolution using diffusion models.
  • Allows for the reuse of the same architecture across different spatial and temporal super-resolution factors.
  • Decomposes the SR task into deterministic and stochastic components to enhance performance.
  • Demonstrates effectiveness on precipitation data, a challenging application in climate science.
Read more
SGD at the Edge of Stability: The Stochastic Sharpness Gap
Fangshuo Liao, Afroditi Kolomvaki, Anastasios Kyrillidis
Optimization Theory
  • Introduces stochastic self-stabilization to explain sharpness suppression in mini-batch SGD.
  • Derives a closed-form expression for the equilibrium sharpness gap that depends on batch size and gradient noise.
  • Demonstrates that smaller batch sizes lead to flatter solutions in the loss landscape.
  • Experimental validation on MLPs, CNNs, and ResNets shows quantitative agreement with theoretical predictions.
Read more
Temporal Taskification in Streaming Continual Learning: A Source of Evaluation Instability
Nicolae Filat, Ahmed Hussain, Konstantinos Kalogiannis, Elena Burceanu
Time Series
  • Temporal taskification is a structural component of evaluation in streaming CL.
  • Different valid splits of the same data stream can lead to varying CL regimes and performance metrics.
  • The proposed framework allows for efficient diagnosis of taskification robustness before model training.
  • Shorter taskifications result in noisier patterns and greater sensitivity to boundary perturbations.
Read more
Frequency-Forcing: From Scaling-as-Time to Soft Frequency Guidance
Weitao Du
Generative Models Computer Vision
  • Frequency-Forcing combines hard and soft frequency guidance for improved image generation.
  • The method utilizes a self-forcing signal derived from data, avoiding reliance on pretrained models.
  • Frequency-Forcing consistently outperforms strong baselines in FID scores on ImageNet-256.
  • The approach maintains compatibility with standard flow-matching architectures.
Read more
Task-specific Subnetwork Discovery in Reinforcement Learning for Autonomous Underwater Navigation
Yi-Ling Liu, Melvin Laux, Mariela De Lucas Alvarez, Frank Kirchner, Rebecca Adam
Reinforcement Learning Robotics Interpretability
  • MTRL effectively utilizes shared knowledge across tasks, indicating successful knowledge sharing.
  • Only a small fraction of network weights are task-specific, suggesting minimal specialization is needed for individual objectives.
  • Context variables play a crucial role in enabling the network to differentiate between related tasks.
Read more
TabSHAP
Aryan Chaudhary, Prateek Agarwal, Tejasvi Alladi
Large Language Models Interpretability
  • Introduces TabSHAP, a model-agnostic interpretability framework for LLM-based tabular classifiers.
  • Utilizes Jensen-Shannon divergence for distributional attribution, capturing shifts in model confidence.
  • Implements feature-level atomic masking to maintain prompt syntax and semantic integrity.
  • Demonstrates significantly higher faithfulness in feature attribution compared to existing methods.
Read more