AI-generated summaries

Today's ML research,
without the noise.

Daily summaries of the latest machine learning papers from arXiv, processed every 8 hours.

40 Papers today
8h Update frequency
7 Days of history
Unsupervised Learning of Inter-Object Relationships via Group Homomorphism
Kyotaro Ushida, Takayuki Komatsu, Yoshiyuki Ohmura, Yasuo Kuniyoshi
Computer Vision Theory Robotics
  • Introduces an unsupervised learning method based on group homomorphism to model inter-object relationships.
  • Demonstrates the ability to segment multiple objects and extract motion laws without ground-truth labels.
  • Highlights the importance of algebraic geometric constraints for creating interpretable representations.
  • Aims to replicate cognitive development processes observed in preverbal infants.
Read more
Task-specific Subnetwork Discovery in Reinforcement Learning for Autonomous Underwater Navigation
Yi-Ling Liu, Melvin Laux, Mariela De Lucas Alvarez, Frank Kirchner, Rebecca Adam
Reinforcement Learning Robotics Interpretability
  • MTRL effectively utilizes shared knowledge across tasks, indicating successful knowledge sharing.
  • Only a small fraction of network weights are task-specific, suggesting minimal specialization needed for individual tasks.
  • Context variables play a crucial role in enabling task differentiation in MTRL.
Read more
Quotient-Space Diffusion Models
Yixian Xu, Yusong Wang, Shengjie Luo, Kaiyuan Gao, Tianyu He, Di He, Chang Liu
Generative Models
  • Introduces a formal framework for diffusion modeling on quotient spaces.
  • Simplifies learning by treating equivalent objects as a single entity.
  • Reduces the necessity of learning group actions, enhancing model efficiency.
  • Empirical results show improved performance in molecular structure generation.
Read more
Even More Guarantees for Variational Inference in the Presence of Symmetries
Lena Zellinger, Antonio Vergari
Theory Optimization
  • Derives sufficient conditions for exact recovery of the mean using FKL and α-divergences.
  • Extends previous results on robust variational inference under target symmetries.
  • Provides guidelines for choosing appropriate variational families based on theoretical insights.
  • Highlights potential optimization failures when sufficient conditions are not met.
Read more
A Green-Integral-Constrained Neural Solver with Stochastic Physics-Informed Regularization
Mohammad Mahdi Abedi, David Pardo, Tariq Alkhalifah
Theory Efficient ML Optimization
  • Introduction of a Green-Integral neural solver for the Helmholtz equation.
  • Elimination of second-order spatial derivatives and boundary layers through integral representation.
  • Significant reduction in computational cost and training time via FFT-based convolution.
  • Hybrid GI+PDE loss improves accuracy in strong scattering regions.
Read more
Transferable Physics-Informed Representations via Closed-Form Head Adaptation
Jian Cheng Wong, Isaac Yin Chung Lai, Pao-Hsiung Chiu, Chin Chun Ooi, Abhishek Gupta, Yew-Soon Ong
Theory Optimization Efficient ML
  • Introduction of Pi-PINN, a framework for transferable physics-informed representations.
  • Utilization of closed-form head adaptation to reduce computational costs in adapting to new PDE instances.
  • Demonstrated synergy between data-driven multi-task learning and physics-informed losses.
  • Empirical results show significant speed and accuracy improvements over traditional PINNs.
Read more
A Deep U-Net Framework for Flood Hazard Mapping Using Hydraulic Simulations of the Wupper Catchment
Christian Lammers, Fernando Arévalo, Leonie Märker-Neuhaus, Daniel Heinenberg, Christian Förster, Karl-Heinz Spies
Efficient ML
  • Development of a deep learning surrogate model for flood prediction using U-Net architecture.
  • Significant reduction in computation time compared to traditional hydraulic simulations.
  • Validation of the model using real hydraulic simulation data from the Wupper catchment.
  • Demonstration of the model's ability to generalize across different topographies.
Read more
Preconditioned DeltaNet: Curvature-aware Sequence Modeling for Linear Recurrences
Neehal Tumma, Noel Loo, Daniela Rus
NLP Large Language Models Efficient ML
  • Introduction of preconditioning to delta-rule recurrences enhances optimization by accounting for curvature in least-squares loss.
  • Derivation of equivalences between linear attention and DeltaNet in the context of preconditioning.
  • Development of efficient chunkwise parallel algorithms for Preconditioned Linear Attention (PLA) and Preconditioned DeltaNet (PDN).
  • Empirical improvements in performance on synthetic tasks and language modeling at large scales.
Read more
Temporal Taskification in Streaming Continual Learning: A Source of Evaluation Instability
Nicolae Filat, Ahmed Hussain, Konstantinos Kalogiannis, Elena Burceanu
Time Series
  • Temporal taskification is a structural component of evaluation in streaming CL.
  • Different valid splits of the same data stream can lead to different CL regimes.
  • The proposed framework allows for efficient diagnosis of taskification robustness before training.
  • Shorter taskifications result in noisier patterns and greater sensitivity to boundary perturbations.
Read more
Fairness under uncertainty in sequential decisions
Michelle Seng Ah Lee, Kirtan Padh, David Watson, Niki Kilbertus, Jatinder Singh
Reinforcement Learning Theory Optimization
  • Introduces a taxonomy of uncertainty in sequential decision-making.
  • Formalizes model and feedback uncertainty using counterfactual logic and reinforcement learning.
  • Demonstrates the potential harms of ignoring unobserved outcomes in decision-making.
  • Shows that uncertainty-aware exploration can improve fairness metrics.
Read more
The Path Not Taken: Duality in Reasoning about Program Execution
Eshgin Hasanov, Md Mahadi Hassan Sibat, Santu Karmaker, Aashish Yadavally
Large Language Models
  • Current benchmarks for LLMs in program execution are limited and prone to data contamination.
  • The concept of duality in reasoning about program execution introduces forward and backward reasoning tasks.
  • DEXBENCH, the proposed benchmark, evaluates LLMs on both execution and counterfactual reasoning.
  • Dual-path reasoning provides a more reliable measure of causal understanding in program execution.
Read more
TRAVELFRAUDBENCH: A Configurable Evaluation Framework for GNN Fraud Ring Detection in Travel Networks
Bhavana Sajja
Graph Learning
  • TFG introduces a configurable framework for evaluating GNNs in fraud detection specific to travel networks.
  • The framework simulates three distinct fraud ring types with a heterogeneous graph structure.
  • GraphSAGE and RGCN-proj significantly outperform traditional MLP methods in fraud detection tasks.
  • Detection capabilities vary across different fraud ring topologies, highlighting the need for tailored approaches.
Read more
Interpretable Quantile Regression by Optimal Decision Trees
Valentin Lemaire, Gaël Aglin, Siegfried Nijssen
Interpretability
  • Introduces Quantile DL8.5 (QDL8.5) for optimal quantile regression trees.
  • Provides predictions for the complete conditional distribution without prior distribution assumptions.
  • Enhances interpretability and robustness by learning multiple trees for different quantiles.
  • Achieves high accuracy with minimal computational overhead compared to traditional methods.
Read more
Droplet-LNO: Physics-Informed Laplace Neural Operators for Accurate Prediction of Droplet Spreading Dynamics on Complex Surfaces
Ganesh Sahadeo Meshram, Partha Pratim Chakrabarti, Suman Chakraborty
Theory Efficient ML Optimization
  • Introduction of PI-LNO for modeling droplet dynamics on complex surfaces.
  • Significant reduction in computational time compared to traditional CFD methods.
  • Outperforms existing state-of-the-art models in accuracy and efficiency.
  • Utilizes a physics-regularized loss function to ensure physically feasible predictions.
Read more
HARBOR: Automated Harness Optimization
Biswa Sengupta, Jinhua Wang
NLP Large Language Models Optimization
  • Harness design is a critical aspect of deploying long-horizon language models, often overshadowing the model itself.
  • HARBOR formalizes Automated Harness Optimization as a constrained noisy optimization problem.
  • A case study reveals that manual tuning is often ineffective, with only one out of four rounds achieving a statistically significant improvement.
  • The paper advocates for treating harness tuning as a hyper-parameter optimization problem.
Read more
Dilated CNNs for Periodic Signal Processing: A Low-Complexity Approach
Eli Gildish, Michael Grebshtein, Igor Makienko
Time Series Efficient ML Audio & Speech
  • R-DCNN offers a low-complexity solution for denoising periodic signals.
  • The method requires only a single observation for training, enhancing efficiency.
  • R-DCNN can generalize across signals with varying frequencies through resampling.
  • Performance is comparable to classical autoregressive methods and conventional DCNNs.
Read more
FairyFuse: Multiplication-Free LLM Inference on CPUs via Fused Ternary Kernels
Fei Zuo, Xiaoyan Xi, Quanyi Zeng, Feiyu Wang, Ho Fai Leung
NLP Large Language Models Efficient ML
  • FairyFuse is the first ternary-weight GEMV kernel on x86 CPUs that eliminates all floating-point multiplications.
  • The system consolidates eight sub-GEMVs into a single SIMD-friendly loop, achieving a 1.55× speedup over unfused execution.
  • Ternary packing shifts computational efficiency from GPUs to CPUs, making CPUs the preferred target for extreme quantization.
  • FairyFuse achieves competitive throughput with 4-bit baselines while utilizing only 2-bit storage.
Read more
ARFBench: Benchmarking Time Series Question Answering Ability for Software Incident Response
Stephan Xie, Ben Cohen, Mononito Goswami, Junhong Shen, Emaad Khwaja, Chenghao Liu, David Asker, Othmane Abou-Amal, Ameet Talwalkar
Time Series Multimodal Large Language Models
  • ARFBench is the first benchmark specifically designed for evaluating TSQA in software incident response.
  • Frontier VLMs outperform existing models, with GPT-5 achieving notable accuracy and F1 scores.
  • Hybrid TSFM-VLM models show promise, achieving performance comparable to leading models.
  • A model-expert oracle demonstrates complementary strengths, establishing a new superhuman frontier for TSQA.
Read more
Sink-Token-Aware Pruning for Fine-Grained Video Understanding in Efficient Video LLMs
Kibum Kim, Jiwan Kim, Kyle Min, Yueqi Wang, Jinyoung Moon, Julian McAuley, Chanyoung Park
Computer Vision Large Language Models Efficient ML
  • Existing visual token pruning methods are inadequate for fine-grained video understanding tasks.
  • Sink tokens significantly hinder model performance by distorting visual evidence.
  • SToP introduces a sink score to effectively prune semantically uninformative tokens.
  • The method shows substantial performance improvements across diverse benchmarks.
Read more
Domain-Aware Hierarchical Contrastive Learning for Semi-Supervised Generalization Fault Diagnosis
Junyu Ren, Wensheng Gan, Philip S Yu
Time Series
  • Introduces DAHCL framework to improve fault diagnosis under unseen conditions.
  • Addresses pseudo-label bias by incorporating domain-specific geometric characteristics.
  • Utilizes uncertain samples effectively through fuzzy contrastive supervision.
  • Evaluates performance under realistic noisy conditions, enhancing practical applicability.
Read more
IRIS: Interpolative Rényi Iterative Self-play for Large Language Model Fine-Tuning
Wenjie Liao, Like Wu, Liangjie Zhao, Shihui Xu, Shigeru Fujimura
NLP Large Language Models Optimization
  • IRIS provides a unified framework for self-play fine-tuning using Rényi divergence.
  • The framework allows for adaptive adjustment of the divergence objective based on training stages.
  • Empirical results show significant performance improvements over existing self-play methods.
  • IRIS achieves competitive results with fewer annotated samples compared to standard supervised fine-tuning.
Read more
A Hybridizable Neural Time Integrator for Stable Autoregressive Forecasting
Brooks Kinch, Xiaozhe Hu, Yilong Huang, Martine Dyring Hansen, Sunniva Meltzer, Nathaniel Donald Hamlin, David Sirajuddin, Eric C. Cyr, Nathaniel Trask
Time Series Theory Efficient ML
  • Introduces a hybrid autoregressive transformer embedded in a mixed finite element framework for stable forecasting.
  • Proves preservation of discrete energies and uniform gradient bounds, avoiding the exploding gradient problem.
  • Achieves a 65× reduction in model parameters while outperforming state-of-the-art models in chaotic system forecasting.
  • Demonstrates a 9,000× speedup in real-time simulations for a fusion component using only 12 training simulations.
Read more
Forget, Then Recall: Learnable Compression and Selective Unfolding via Gist Sparse Attention
Yuzhen Mao, Michael Y. Li, Emily B. Fox
NLP Large Language Models Efficient ML
  • Introduces Gist Sparse Attention (GSA) for efficient long-context modeling in LLMs.
  • Combines learnable compression with selective unfolding to improve attention mechanisms.
  • Achieves significant performance improvements over existing compression and sparse attention methods.
  • Enables multi-resolution context access with reduced computational complexity.
Read more
An effective variant of the Hartigan $k$-means algorithm
François Clément, Stefan Steinerberger
Optimization Theory Efficient ML
  • Smartigan improves upon Hartigan's k-means algorithm by an additional 2-5%.
  • The algorithm encourages exploration of the clustering space, particularly beneficial in high-dimensional settings.
  • Smartigan-stability provides a strong guarantee for cluster assignments, enhancing the robustness of the clustering process.
  • Empirical results confirm the superiority of Smartigan over traditional Lloyd's and Hartigan's algorithms.
Read more
Probably Approximately Consensus: On the Learning Theory of Finding Common Ground
Carter Blair, Ben Armstrong, Shiri Alouf-Heffetz, Nimrod Talmon, Davide Grossi
Theory Optimization
  • Introduces a formal definition for passive 1D interval-based consensus finding.
  • Develops an efficient Empirical Risk Minimization (ERM) algorithm.
  • Establishes PAC learning guarantees, including sample complexity bounds.
  • Demonstrates the effectiveness of selective querying strategies in reducing query numbers.
Read more
Geometric Characterisation and Structured Trajectory Surrogates for Clinical Dataset Condensation
Pafue Christy Nganjimi, Andrew Soltan, Danielle Belgrave, Lei Clifton, David Clifton, Anshul Thakur
Theory Efficient ML Optimization
  • Introduces a geometric characterization of trajectory matching in dataset condensation.
  • Identifies a representability bottleneck in traditional trajectory matching methods.
  • Proposes Bézier Trajectory Matching (BTM) to improve the efficiency of dataset condensation.
  • Demonstrates that BTM outperforms standard trajectory matching in clinical datasets, especially in challenging settings.
Read more
Generalizing Numerical Reasoning in Table Data through Operation Sketches and Self-Supervised Learning
Hanjun Cho, Gahyun Yoo, Hanseong Kim, Jay-Yoon Lee
NLP Large Language Models Efficient ML
  • TaNOS framework improves numerical reasoning robustness by addressing reasoning inefficiency, data scarcity, and header dependency.
  • Operation sketches help models focus on contextual reasoning rather than surface-level arithmetic.
  • Self-supervised learning allows for the construction of program-question pairs without manual annotation, enhancing data efficiency.
  • Header anonymization reduces reliance on specific lexical cues, promoting better schema generalization.
Read more
Low-Rank Adaptation Redux for Large Models
Bingcong Li, Yilang Zhang, Georgios B. Giannakis
Large Language Models Optimization Efficient ML
  • LoRA is a leading method for parameter-efficient fine-tuning of large models, significantly reducing computational and memory costs.
  • The paper categorizes advancements in LoRA into architectural design, efficient optimization, and diverse applications.
  • Classical signal processing tools provide valuable insights for improving LoRA methods and addressing challenges in deep learning.
  • The authors advocate for a systematic approach to LoRA design, informed by first-principles guidelines from signal processing.
Read more
Absorber LLM: Harnessing Causal Synchronization for Test-Time Training
Zhixin Zhang, Shabo Zhang, Chengcan Wu, Zeming Wei, Meng Sun
Large Language Models Efficient ML NLP
  • Absorber LLM preserves causal relationships between historical contexts and future inferences.
  • The method optimizes context absorption through self-supervised causal synchronization.
  • Absorber LLM outperforms traditional transformers and prior parameter memory methods in both efficiency and accuracy.
  • The approach enables scalable inference in real-world applications involving long-context data.
Read more
Graph Neural Network-Informed Predictive Flows for Faster Ford-Fulkerson and PAC-Learnability
Eleanor Wiesler, Trace Baxley
Graph Learning Optimization Computer Vision
  • Integration of GNNs with the Ford-Fulkerson algorithm to improve max-flow computation speed.
  • Introduction of a Message Passing Graph Neural Network (MPGNN) that learns edge importance probabilities.
  • Development of a modified Ford-Fulkerson procedure that prioritizes high-value augmenting paths.
  • Theoretical framework connecting prediction quality to algorithmic efficiency.
Read more
Validating a Deep Learning Algorithm to Identify Patients with Glaucoma using Systemic Electronic Health Records
John Xiang, Rohith Ravindranath, Sophia Y. Wang
Efficient ML
  • The study demonstrates the transportability of a pretrained deep learning model for glaucoma risk assessment using EHR data.
  • The model achieved an AUROC of 0.883 and a PPV of 0.657, indicating effective identification of glaucoma patients.
  • Calibration of the model predictions aligned with clinical outcomes, enhancing its potential for practical application.
  • Fine-tuning the model on local data improved its performance, highlighting the importance of adapting models to specific health systems.
Read more
Measure Twice, Click Once: Co-evolving Proposer and Visual Critic via Reinforcement Learning for GUI Grounding
Wenkai Wang, Xiyun Li, Hongcan Guo, Wenhao Yu, Tianqing Fang, Haitao Mi, Dong Yu, Shengyu Zhang
Computer Vision Reinforcement Learning Multimodal
  • Introduction of a Propose-then-Critic framework for GUI grounding.
  • Utilization of a co-evolutionary reinforcement learning strategy to enhance model capabilities.
  • Dynamic maturity mechanism to balance prediction accuracy and candidate diversity.
  • Significant improvements in grounding accuracy and critic reliability.
Read more
Fine-Tuning Regimes Define Distinct Continual Learning Problems
Paul-Tiberiu Iordache, Elena Burceanu
Theory Optimization
  • Fine-tuning regimes are crucial evaluation variables in continual learning.
  • Changing the trainable depth affects the optimization geometry and update signals.
  • Empirical results show that method rankings can vary significantly across different regimes.
  • Deeper adaptation regimes correlate with higher forgetting and larger update magnitudes.
Read more
Transferable SCF-Acceleration through Solver-Aligned Initialization Learning
Eike S. Eberhard, Viktor Kotsev, Timm Güthle, Stephan Günnemann
Optimization Efficient ML Theory
  • SAIL improves the quality of initial guesses for SCF solvers by training on solver dynamics.
  • The Effective Relative Iteration Count (ERIC) is introduced as a more accurate performance metric.
  • SAIL achieves significant reductions in ERIC across various molecular sizes, outperforming previous methods.
  • The method extends machine learning SCF acceleration to larger drug-like molecules, enhancing computational efficiency.
Read more
Early Detection of Latent Microstructure Regimes in Limit Order Books
Prakul Sunil Hiremath, Vruksha Arun Hiremath
Time Series Theory
  • Introduces a causal regime model for LOBs with identifiable latent build-up phases.
  • Derives theoretical guarantees for early detection of stress onset.
  • Proposes a novel trigger-based detector with formal foundations.
  • Demonstrates superior performance over traditional detection methods in simulations.
Read more
MCAP: Deployment-Time Layer Profiling for Memory-Constrained LLM Inference
Anurita Das
Large Language Models Efficient ML NLP
  • MCAP enables load-time per-layer precision decisions, enhancing flexibility in model deployment.
  • The method significantly increases decode throughput compared to existing systems.
  • NVE can run larger models in constrained memory environments without performance loss.
  • MCAP provides a single runtime signal that couples precision routing and memory placement decisions.
Read more
Relocation of compact sets in $ ext{R}^n$ by diffeomorphisms and linear separability of datasets in $ ext{R}^n$
Xiao-Song Yang, Xuan Zhou, Qi Zhou
Theory
  • Establishes a theory for relocating compact sets in R^n using diffeomorphisms.
  • Proves that a differentiable embedding exists to make compact datasets linearly separable in R^(n+1).
  • Demonstrates that width-n deep neural networks can achieve linear separability for compact datasets.
  • Connects concepts from differential topology with applications in deep learning.
Read more
PrismaDV: Automated Task-Aware Data Unit Test Generation
Hao Chen, Arnab Phani, Sebastian Schelter
Theory Efficient ML
  • PrismaDV generates task-aware data unit tests by analyzing downstream task code and dataset profiles.
  • The SIFTA framework allows for continuous adaptation of data unit tests based on execution outcomes.
  • PrismaDV outperforms existing task-agnostic and task-aware frameworks in generating relevant unit tests.
  • The system addresses common shortcomings in data unit testing, such as manual maintenance and partial coverage.
Read more
A Scale-Adaptive Framework for Joint Spatiotemporal Super-Resolution with Diffusion Models
Max Defez, Filippo Quarenghi, Mathieu Vrac, Stephan Mandt, Tom Beucler
Generative Models Time Series Computer Vision
  • Introduces a scale-adaptive framework for joint spatiotemporal super-resolution using diffusion models.
  • Decomposes spatiotemporal SR into deterministic and stochastic components, enhancing model flexibility.
  • Requires only three retuned hyperparameters for adaptation across different SR factors.
  • Demonstrated effectiveness on precipitation data, supporting applications in climate science.
Read more
JEPAMatch: Geometric Representation Shaping for Semi-Supervised Learning
Ali Aghababaei-Harandi, Aude Sportisse, Massih-Reza Amini
Computer Vision Theory Efficient ML
  • JEPAMatch addresses class imbalance and slow convergence issues in semi-supervised learning.
  • The method integrates geometric representation shaping inspired by LeJEPA into the training process.
  • Extensive experiments show significant performance improvements over existing methods.
  • The approach can be adapted to various FixMatch variants, enhancing its applicability.
Read more