AI-generated summaries

Today's ML research,
without the noise.

Daily summaries of the latest machine learning papers from arXiv, processed every 8 hours.

59 Papers today
8h Update frequency
7 Days of history
Off-Policy Learning with Limited Supply
Koichi Tanaka, Ren Kishimoto, Bushun Kawagishi, Yusuke Narita, Yasuo Yamamoto, Nobuyuki Shimizu, Yuta Saito
Reinforcement Learning Theory Optimization
  • Conventional greedy OPL methods are suboptimal in limited supply scenarios.
  • The paper introduces OPLS, which focuses on relative expected rewards for better item allocation.
  • Theoretical analysis proves the existence of superior policies in limited supply settings.
  • OPLS does not incur additional computational costs compared to existing methods.
Read more
SOL-ExecBench: Speed-of-Light Benchmarking for Real-World GPU Kernels Against Hardware Limits
Edward Lin, Sahil Modi, Siva Kumar Sastry Hari, Qijing Huang, Zhifan Ye, Nestor Qin, Fengzhe Zhou, Yuan Zhang, Jingquan Wang, Sana Damani, Dheeraj Peri, Ouye Xie, Aditya Kane, Moshe Maor, Michael Behar, Triston Cao, Rishabh Mehta, Vartika Singh, Vikram Sharma Mailthody, Terry Chen, Zihao Ye, Hanfeng Chen, Tianqi Chen, Vinod Grover, Wei Chen, Wei Liu, Eric Chung, Luis Ceze, Roger Bringmann, Cyril Zeller, Michael Lightstone, Christos Kozyrakis, Humphrey Shi
Optimization Efficient ML Generative Models
  • SOL-ExecBench benchmarks GPU kernels against hardware Speed-of-Light limits rather than software baselines.
  • The framework includes 235 optimization problems from diverse AI models, ensuring broad applicability.
  • SOL Score quantifies performance improvements relative to analytically derived hardware limits.
  • A sandboxed evaluation harness enhances reliability and prevents reward-hacking in kernel optimization.
Read more
Engineering Verifiable Modularity in Transformers via Per-Layer Supervision
J. Clayton Kerce
Interpretability
  • Introduces per-layer supervision to enhance modularity in transformer models.
  • Demonstrates that per-layer supervision leads to significantly larger ablation effects.
  • Establishes a methodology for transforming interpretability into active control.
  • Validates findings through engineered features and causal experiments.
Read more
Enhancing the Parameterization of Reservoir Properties for Data Assimilation Using Deep VAE-GAN
Marcio Augusto Sampaio, Paulo Henrique Ranazzi, Martin Julian Blunt
Generative Models
  • Introduces VAE-GAN to enhance parameterization in reservoir data assimilation.
  • Addresses limitations of traditional ensemble methods in handling non-Gaussian distributions.
  • Demonstrates improved geological plausibility and history matching in reservoir simulations.
  • Validates methodology through two distinct case studies with categorical and continuous data.
Read more
STEP: Scientific Time-Series Encoder Pretraining via Cross-Domain Distillation
Chen Zhang, Liwei Liu, Jun Tao, Xiaoyu Yang, Xuenan Xu, Kai Chen, Bowen Zhou, Wen Wu, Chao Zhang
Time Series
  • STEP framework leverages cross-domain distillation to enhance scientific time-series representation learning.
  • Introduces adaptive patching and statistics compensation to handle diverse and heterogeneous scientific signals.
  • Demonstrates the transferability of knowledge from foundation models in related domains to improve performance on scientific tasks.
  • Achieves strong performance across various scientific time series tasks, indicating its effectiveness as a pretraining paradigm.
Read more
NANOZK: Layerwise Zero-Knowledge Proofs for Verifiable Large Language Model Inference
Zhaohui Geoffrey Wang
Large Language Models Theory Efficient ML
  • NANOZK provides a cryptographic verification mechanism for LLM inference, addressing trust issues in LLM-as-a-service.
  • The layerwise proof framework allows for independent layer proofs, significantly reducing computational overhead.
  • Lookup table approximations for non-arithmetic operations ensure zero accuracy loss during verification.
  • Fisher information is used to prioritize layer verification, enhancing efficiency in resource-constrained scenarios.
Read more
Position: Spectral GNNs Are Neither Spectral Nor Superior for Node Classification
Qin Jiang, Chengjia Wang, Michael Lones, Dongdong Chen, Wei Pang
Graph Learning Theory
  • Spectral GNNs do not utilize true Fourier bases for graph signals.
  • Polynomial approximations in Spectral GNNs are theoretically flawed.
  • The performance of GCNs is attributed to message-passing dynamics rather than spectral filtering.
  • Empirical success of models like MagNet and HoloNet is linked to implementation issues, not spectral properties.
Read more
Discovering What You Can Control: Interventional Boundary Discovery for Reinforcement Learning
Jiaxin Liu
Reinforcement Learning Robotics Theory
  • IBD formalizes the Causal Sphere of Influence, distinguishing between causally relevant and confounded dimensions.
  • The method requires no learned models and can be applied as a preprocessing step to any RL algorithm.
  • Empirical results show that IBD outperforms traditional observational feature selection, especially in high distractor scenarios.
  • IBD provides a diagnostic framework to decompose environmental difficulties into representational confusion and exploration challenges.
Read more
FlowMS: Flow Matching for De Novo Structure Elucidation from Mass Spectra
Jianan Nie, Peng Gao
Generative Models Graph Learning
  • FlowMS is the first discrete flow matching framework for de novo molecular generation from mass spectra.
  • It achieves state-of-the-art performance on 5 out of 6 metrics on the NPLIB1 benchmark.
  • FlowMS demonstrates a 9.15% top-1 accuracy, surpassing previous methods.
  • The framework effectively enforces chemical formula constraints during molecular generation.
Read more
Transformers Learn Robust In-Context Regression under Distributional Uncertainty
Hoang T. H. Cao, Hai D. V. Trinh, Tho Quan, Lan V. Truong
Theory
  • Transformers can perform in-context learning for linear regression under distributional uncertainties.
  • The study systematically evaluates the performance of Transformers against classical regression methods under various distributional shifts.
  • Transformers demonstrate robustness and adaptability beyond traditional estimators in non-Gaussian and non-i.i.d. settings.
  • The research identifies specific regimes where Transformers outperform classical methods, including sample size and feature covariance.
Read more
VC-Soup: Value-Consistency Guided Multi-Value Alignment for Large Language Models
Hefei Xu, Le Wu, Yu Wang, Min Hou, Han Wu, Zhen Zhang, Meng Wang
NLP Large Language Models Reinforcement Learning
  • VC-Soup addresses the limitations of existing multi-value alignment methods for LLMs.
  • The framework introduces a value consistency metric to filter low-consistency data.
  • It trains value-consistent policy models that enhance linear mode connectivity.
  • The approach combines policies and applies Pareto filtering for balanced performance.
Read more
Beyond Passive Aggregation: Active Auditing and Topology-Aware Defense in Decentralized Federated Learning
Sheng Pan, Niansheng Tang
Federated Learning Graph Learning Theory
  • Introduction of an active auditing framework for DFL to counter adaptive backdoor attacks.
  • Development of three novel auditing metrics to expose hidden backdoors in local models.
  • Implementation of a topology-aware defense placement strategy to enhance resilience.
  • Theoretical analysis of convergence rates under co-evolving attack and defense dynamics.
Read more
Are complicated loss functions necessary for teaching LLMs to reason?
Gabriele Carrino, Andrea Sassella, Nicolo Brunello, Federico Toschi, Mark James Carman
NLP Large Language Models Reinforcement Learning
  • Negative feedback is essential for effective learning in LLMs.
  • PPO-style constraints are not necessary for improving mathematical reasoning.
  • RGRA, a simplified variant of GRPO, shows superior performance on reasoning tasks.
  • Simpler reinforcement learning methods can enhance reasoning in LLMs.
Read more
BoundAD: Boundary-Aware Negative Generation for Time Series Anomaly Detection
Xiancheng Wang, Lin Wang, Zhibo Zhang, Rui Wang, Minghang Zhao
Time Series Reinforcement Learning Optimization
  • Introduces a reconstruction-driven framework for generating hard negatives in TSAD.
  • Utilizes reinforcement learning to adaptively optimize the negative sample generation process.
  • Improves anomaly representation learning by focusing on boundary-aware negative samples.
  • Achieves competitive detection performance on benchmark datasets.
Read more
Approximate Subgraph Matching with Neural Graph Representations and Reinforcement Learning
Kaiyang Li, Shihao Ji, Zhipeng Cai, Wei Li
Reinforcement Learning Graph Learning Optimization
  • RL-ASM is the first approach to apply reinforcement learning for high-quality approximate subgraph matching.
  • The method utilizes a Graph Transformer to fully exploit graph information for improved matching.
  • RL-ASM optimizes node pair selection based on long-term rewards rather than greedy heuristics.
  • Extensive experiments show RL-ASM's superior performance compared to traditional ASM algorithms.
Read more
Towards Differentiating Between Failures and Domain Shifts in Industrial Data Streams
Natalia Wojak-Strzelecka, Szymon Bobek, Grzegorz J. Nalepa, Jerzy Stefanowski
Time Series Interpretability
  • The proposed method effectively distinguishes between failures and healthy domain shifts in industrial data streams.
  • Integration of a modified Page-Hinkley changepoint detector enhances the identification of changes in data distribution.
  • Supervised domain-adaptation algorithms facilitate fast online anomaly detection.
  • Explainable AI components support human operators in decision-making processes.
Read more
MST-Direct: Matching via Sinkhorn Transport for Multivariate Geostatistical Simulation with Complex Non-Linear Dependencies
Tchalies Bachmann Schmitz
Optimization Theory Generative Models
  • MST-Direct preserves complex non-linear dependencies in multivariate geostatistical simulations.
  • The method utilizes Optimal Transport theory and the Sinkhorn algorithm for distribution matching.
  • It processes all variables simultaneously, enhancing computational efficiency.
  • Experimental validation shows 100% shape preservation across diverse relationship types.
Read more
Communication-Efficient and Robust Multi-Modal Federated Learning via Latent-Space Consensus
Mohamed Badi, Chaouki Ben Issaid, Mehdi Bennis
Federated Learning Multimodal
  • Introduction of CoMFed, a novel framework for multi-modal federated learning.
  • Utilization of learnable projection matrices for generating compressed latent representations.
  • Implementation of a robust alignment regularizer based on geometric-median consensus.
  • Demonstration of competitive performance on real-world multi-modal datasets.
Read more
Foundations of SchrΓΆdinger Bridges for Generative Modeling
Sophia Tang
Generative Models Theory Optimization
  • SchrΓΆdinger bridges unify various generative modeling frameworks, including diffusion models and flow matching.
  • The paper develops a mathematical foundation for the SchrΓΆdinger bridge problem, linking it to optimal transport and stochastic control.
  • A comprehensive toolkit for constructing SchrΓΆdinger bridges is introduced, facilitating task-specific computational methods.
  • The framework is applicable to diverse problems, including data translation and single-cell state dynamics modeling.
Read more
Frayed RoPE and Long Inputs: A Geometric Perspective
Davis Wertheimer, Aozhong Zhang, Derrick Liu, Penghang Yin, Naigang Wang
NLP Large Language Models Theory
  • RoPE causes performance issues for long inputs due to the dispersion of key/query clusters.
  • Attention mechanisms create sink tokens that help prevent over-mixing of information.
  • RoPE-ID modifies RoPE by applying high-frequency rotation to a fraction of channels, improving generalization.
  • The proposed method shows strong performance on long-context tasks compared to previous tuning-free approaches.
Read more
SLEA-RL: Step-Level Experience Augmented Reinforcement Learning for Multi-Turn Agentic Training
Prince Zizhuang Wang, Shuli Jiang
Reinforcement Learning Large Language Models Robotics
  • SLEA-RL retrieves experiences at each decision step, enhancing adaptability in multi-turn tasks.
  • The framework includes a self-evolving experience library that maintains quality through score-based admission and rate-limited extraction.
  • Empirical validation shows SLEA-RL outperforms standard RL and experience-augmented methods on multiple benchmarks.
Read more
Self-Tuning Sparse Attention: Multi-Fidelity Hyperparameter Optimization for Transformer Acceleration
Arundhathi Dev, Justin Zhan
NLP Large Language Models Efficient ML
  • AFBS-BO automates the hyperparameter tuning process for sparse attention, eliminating the need for manual grid search.
  • The method achieves a 3.4Γ— speedup in hyperparameter discovery and requires significantly fewer evaluations than traditional methods.
  • Configurations discovered by AFBS-BO outperform existing sparse attention baselines while closely matching the quality of dense attention.
  • The framework enables robust performance across diverse transformer architectures and input distributions.
Read more
Hierarchical Latent Structure Learning through Online Inference
Ines Aitsahalia, Kiyohito Iigaya
Theory Time Series Efficient ML
  • HOLMES model integrates hierarchical representation with online inference for latent structure learning.
  • The model uses a nested Chinese Restaurant Process prior and sequential Monte Carlo methods.
  • HOLMES achieves compact representations that support one-shot transfer to higher-level categories.
  • In simulations, HOLMES improves predictive performance in context-dependent tasks compared to flat models.
Read more
Context Bootstrapped Reinforcement Learning
Saaket Agashe, Jayanth Srinivasa, Gaowen Liu, Ramana Kompella, Xin Eric Wang
Reinforcement Learning Large Language Models
  • CBRL effectively addresses exploration inefficiency in RLVR by using few-shot demonstrations.
  • The method employs a stochastic injection mechanism that reduces assistance over time.
  • CBRL shows consistent performance improvements across diverse tasks and model families.
  • The approach is algorithm-agnostic, yielding gains with different reinforcement learning algorithms.
Read more
Adapting Methods for Domain-Specific Japanese Small LMs: Scale, Architecture, and Quantization
Takato Yasuno
NLP Large Language Models Efficient ML
  • Optimal training scale for domain-specific adaptation is identified as 4,000 samples, balancing underfitting and overfitting.
  • Llama-3 models with Japanese pre-training outperform multilingual models in technical domain tasks.
  • Architecture-specific effects of quantization are documented, with Llama-3 models improving under Q4 quantization while GQA models degrade.
  • A complete reproducible pipeline is provided for practitioners to replicate results on consumer hardware.
Read more
Quotient Geometry and Persistence-Stable Metrics for Swarm Configurations
Mark M. Bailey
Robotics Theory Optimization
  • Introduction of a new metric for comparing swarm configurations that is both persistence-stable and symmetry-invariant.
  • Establishment of a quotient formation space and a corresponding formation matching metric, enhancing the understanding of multi-agent systems.
  • Analysis of the metric geometry reveals compactness and completeness under specific conditions, linking to classical configuration spaces.
  • Identification of symmetry-mismatch and persistence-compression mechanisms that affect the expressivity of signatures.
Read more
OCP: Orthogonal Constrained Projection for Sparse Scaling in Industrial Commodity Recommendation
Chen Sun, Beilin Xu, Boheng Tan, Jiacheng Wang, Yuefeng Sun, Rite Bo, Ying He, Yaqiang Zang, Pinghua Gong
Optimization
  • OCP addresses embedding collapse in sparse scaling by enforcing orthogonality in the projection matrix.
  • The method quantifies representation isotropy using Singular Entropy (SE) to analyze the impact of long-tail sparsity.
  • Empirical results show OCP accelerates loss convergence and improves model scalability.
  • Large-scale deployment on JD.com resulted in a 12.97% increase in UCXR and an 8.9% uplift in GMV.
Read more
SpecForge: A Flexible and Efficient Open-Source Training Framework for Speculative Decoding
Shenggui Li, Chao Wang, Yikai Zhu, Yubo Wang, Fan Yin, Shuai Shi, Yefei Chen, Xiaomin Dong, Qiaoling Chen, Jin Pan, Ji Li, Laixin Xie, Yineng Zhang, Lei Yu, Yonggang Wen, Ivor Tsang, Tianwei Zhang
NLP Large Language Models Efficient ML
  • SpecForge provides a scalable and efficient framework for training speculative decoding models.
  • The framework supports EAGLE-3, enabling significant speed improvements in training and inference.
  • SpecBundle offers a suite of high-quality draft models, addressing the scarcity of effective draft models in the community.
  • The proposed methods lead to substantial reductions in inference latency without compromising output quality.
Read more
MolRGen: A Training and Evaluation Setting for De Novo Molecular Generation with Reasoning Models
Philippe Formont, Maxime Darrin, Ismail Ben Ayed, Pablo Piantanida
Large Language Models Reinforcement Learning Generative Models
  • Introduction of MOLRGEN, a large-scale dataset for de novo molecular generation.
  • Development of a diversity-aware top-k scoring mechanism for evaluating generated molecules.
  • Successful training of a 24B LLM using reinforcement learning for molecular generation.
  • Identification of challenges in exploring the chemical space during molecular generation.
Read more
Epistemic Generative Adversarial Networks
Muhammad Mubashar, Fabio Cuzzolin
Generative Models Theory Interpretability
  • Introduction of Epistemic Generative Adversarial Networks (EGANs) to enhance output diversity in GANs.
  • Utilization of Dempster-Shafer theory for uncertainty quantification in generative processes.
  • Modification of the discriminator to predict belief functions instead of probability distributions.
  • Architectural enhancements to the generator for region-wise uncertainty estimation.
Read more
SINDy-KANs: Sparse identification of non-linear dynamics through Kolmogorov-Arnold networks
Amanda A. Howard, Nicholas Zolman, Bruno Jacob, Steven L. Brunton, Panos Stinis
Theory Interpretability Time Series
  • SINDy-KANs combine the strengths of KANs and SINDy for improved interpretability.
  • The framework allows for symbolic regression at the activation function level.
  • SINDy-KANs facilitate the discovery of sparse, interpretable equations from data.
  • The methodology is validated through experiments on various dynamical systems.
Read more
AcceRL: A Distributed Asynchronous Reinforcement Learning and World Model Framework for Vision-Language-Action Models
Chengxuan Lu, Shukuan Wang, Yanjie Li, Wei Liu, Shiji Jin, Fuyuan Qian, Peiming Li, Baigui Sun, Yang Liu
Reinforcement Learning Multimodal Robotics
  • AcceRL eliminates synchronization barriers in reinforcement learning, maximizing hardware utilization.
  • The framework integrates a trainable world model, allowing for improved sample efficiency by 200x.
  • AcceRL achieves state-of-the-art performance on the LIBERO benchmark across all evaluation categories.
  • The architecture demonstrates super-linear scaling in throughput with increased computational resources.
Read more
ALIGN: Adversarial Learning for Generalizable Speech Neuroprosthesis
Zhanqi Zhang, Shun Li, Bernardo L. Sabatini, Mikio Aoi, Gal Mishne
Audio & Speech
  • ALIGN is designed to improve the generalization of BCIs for speech decoding across different sessions.
  • The framework employs adversarial learning to align session representations and extract invariant features.
  • Evaluation results show significant reductions in phoneme and word error rates compared to existing methods.
  • The approach addresses the challenges posed by nonstationarities in neural recordings, enhancing long-term usability of BCIs.
Read more
AgentDS Technical Report: Benchmarking the Future of Human-AI Collaboration in Domain-Specific Data Science
An Luo, Jin Du, Xun Xian, Robert Specht, Fangqiao Tian, Ganghua Wang, Xuan Bi, Charles Fleming, Ashish Kundu, Jayanth Srinivasa, Mingyi Hong, Rui Zhang, Tianxi Li, Galin Jones, Jie Ding
Multimodal
  • AI agents struggle with domain-specific reasoning and multimodal integration.
  • Human expertise is essential for diagnosing modeling failures and making strategic decisions.
  • Human-AI collaboration outperforms both humans and AI working independently.
  • The benchmark challenges are designed to reward domain-specific insights over generic methods.
Read more
Rigorous Error Certification for Neural PDE Solvers: From Empirical Residuals to Solution Guarantees
Amartya Mukherjee, Maxwell Fitzsimmons, David C. Del Rey FernΓ‘ndez, Jun Liu
Theory
  • Establishes a theoretical link between residual-based training objectives and solution space error for PINNs.
  • Derives generalization bounds that can be computed without access to the true solution.
  • Demonstrates the necessity of structural regularity in addition to residual control for reliable neural PDE solutions.
  • Provides formal verification of error bounds across multiple PDE types, ensuring empirical stability.
Read more
Best-of-Both-Worlds Multi-Dueling Bandits: Unified Algorithms for Stochastic and Adversarial Preferences under Condorcet and Borda Objectives
S. Akash, Pratik Gajane, Jawar Singh
Theory
  • Introduces best-of-both-worlds algorithms for multi-dueling bandits under Condorcet and Borda objectives.
  • MetaDueling achieves optimal pseudo-regret in adversarial and stochastic settings simultaneously.
  • SA-MiDEX adapts from stochastic to adversarial strategies based on observed deviations.
  • Establishes matching lower bounds for the proposed algorithms, confirming their optimality.
Read more
Towards Noise-Resilient Quantum Multi-Armed and Stochastic Linear Bandits
Zhuoyue Chen, Kechao Cai
Theory Optimization
  • Introduction of a noise-robust QMC algorithm (BQMC) that improves estimation accuracy in noisy quantum environments.
  • Development of noise-resilient quantum bandit algorithms (NR-QUCB and NR-QLinUCB) that maintain performance advantages over classical methods.
  • Extensive experimental validation showing improved regret performance under multiple quantum noise models.
Read more
Attack by Unlearning: Unlearning-Induced Adversarial Attacks on Graph Neural Networks
Jiahao Zhang, Yilong Wang, Suhang Wang
Graph Learning Optimization Theory
  • Introduces unlearning corruption attacks that exploit the unlearning process in GNNs.
  • Formulates the attack as a bi-level optimization problem to address black-box unlearning challenges.
  • Demonstrates that carefully designed unlearning requests can lead to significant accuracy degradation.
  • Raises concerns about the robustness of GNNs under real-world regulatory demands.
Read more
From ex(p) to poly: Gaussian Splatting with Polynomial Kernels
Joerg H. Mueller, Martin Winter, Markus Steinberger
Computer Vision Efficient ML
  • Introduction of an N-th-order polynomial kernel approximation for Gaussian Splatting.
  • Compatibility with existing datasets while significantly reducing computational overhead.
  • Demonstrated performance improvements of 4% to 15% with minimal impact on image quality.
  • Formal mathematical derivation proving invariance of anti-aliasing normalization factors.
Read more
Revisiting Label Inference Attacks in Vertical Federated Learning: Why They Are Vulnerable and How to Defend
Yige Liu, Dexuan Xu, Zimai Guo, Yongzhi Cao, Hanpin Wang
Federated Learning Theory
  • Existing assumptions about the effectiveness of bottom models in representing labels are misleading.
  • Mutual information analysis reveals that bottom models focus on feature extraction, while top models handle label mapping.
  • The 'model compensation' phenomenon highlights the vulnerabilities of LIAs in VFL.
  • A novel defense technique involving cut layer adjustment significantly reduces LIA attack accuracy.
Read more
Automatic Configuration of LLM Post-Training Pipelines
Channe Chwa, Xinle Wu, Yao Lu
Large Language Models Reinforcement Learning Optimization
  • AutoPipe is a budget-aware framework for LLM post-training configuration selection.
  • It employs a dataset-conditioned ranking surrogate to provide transferable guidance across datasets.
  • The framework adapts online using Bayesian optimization and a Gaussian-process residual surrogate.
  • Early-stopping mechanisms are implemented to minimize evaluation costs.
Read more
Robustness, Cost, and Attack-Surface Concentration in Phishing Detection
Julian Allagan, Mohamed Elbakary, Zohreh Safari, Weizheng Gao, Gabrielle Morgan, Essence Morgan, Vladimir Deriglazov
Theory Optimization
  • High accuracy in phishing detection does not guarantee robustness against feature manipulation.
  • The proposed cost-aware evasion framework reveals critical insights into the economics of feature edits.
  • Robustness is significantly influenced by a small number of low-cost features, highlighting the need for strategic feature selection.
  • The concept of action-set-limited invariance indicates that improving robustness requires changes to feature representation or cost models.
Read more
MLOW: Interpretable Low-Rank Frequency Magnitude Decomposition of Multiple Effects for Time Series Forecasting
Runze Yang, Longbing Cao, Xiaoming Wu, Xin You, Kun Fang, Jianxun Li, Jie Yang
Time Series
  • MLOW provides a novel frequency-based decomposition approach for time series forecasting.
  • Introduces Hyperplane-NMF to enhance interpretability and efficiency in low-rank decomposition.
  • Demonstrates robustness to noise and effective separation of multiple effects in time series.
  • Allows for flexible selection of input horizons and frequency levels to address spectral leakage.
Read more
Data-efficient pre-training by scaling synthetic megadocs
Konwoo Kim, Suhas Kotha, Yejin Choi, Tatsunori Hashimoto, Nick Haber, Percy Liang
NLP Large Language Models Efficient ML
  • Synthetic data augmentation can significantly improve pre-training efficiency in language models.
  • The introduction of 'megadocs' enhances data efficiency and model performance.
  • Optimal mixing and epoching strategies are crucial for leveraging synthetic data effectively.
  • The study demonstrates that improvements in loss scaling are more pronounced with increased synthetic data generation.
Read more
BVSIMC: Bayesian Variable Selection-Guided Inductive Matrix Completion for Improved and Interpretable Drug Discovery
Sijian Fan, Liyan Xiong, Dayuan Wang, Guoshuai Cai, Ray Bai
Interpretability
  • BVSIMC improves predictive accuracy and interpretability in drug discovery by incorporating variable selection from side features.
  • The model utilizes spike-and-slab priors to filter out irrelevant or noisy side information.
  • BVSIMC outperforms several existing methods in both synthetic and real-world drug discovery applications.
  • The approach reveals clinically meaningful side features that can guide drug development.
Read more
Improving RCT-Based Treatment Effect Estimation Under Covariate Mismatch via Calibrated Alignment
Amir Asiaee, Samhita Pal
Theory
  • Introduction of CALM framework for embedding alignment in treatment effect estimation.
  • Derivation of finite-sample risk bounds that clarify when embedding alignment is superior to imputation.
  • Demonstration of CALM's effectiveness through extensive simulations, particularly in nonlinear settings.
Read more
Fundamental Limits of Neural Network Sparsification: Evidence from Catastrophic Interpretability Collapse
Dip Roy, Rajiv Misra, Sanjay Kumar Singh
Interpretability
  • Extreme sparsification leads to a significant collapse in local feature interpretability despite stable global representation quality.
  • The phenomenon of interpretability collapse is intrinsic to the sparsification process, not limited to specific algorithms or training durations.
  • The collapse scales with dataset complexity, indicating more severe interpretability issues for complex real-world data.
Read more
SHAPCA: Consistent and Interpretable Explanations for Machine Learning Models on Spectroscopy Data
Mingxing Zhang, Nicola Rossberg, Simone Innocente, Katarzyna Komolibus, Rekha Gautam, Barry O'Sullivan, Luca Longo, Andrea Visentin
Interpretability
  • SHAPCA combines PCA and SHAP to improve interpretability of machine learning models on spectroscopy data.
  • The framework allows for explanations in the original input space, enhancing practical applicability.
  • SHAPCA provides both global and local perspectives on model predictions.
  • The method demonstrates improved consistency and stability in feature importance across training runs.
Read more
Lightweight Adaptation for LLM-based Technical Service Agent: Latent Logic Augmentation and Robust Noise Reduction
Yi Yu, Junzhuo Ma, Chenghuang Shen, Xingyan Liu, Jing Gu, Hangyi Sun, Guangquan Hu, Jianfeng Liu, Weiting Liu, Mingyue Pu, Yu Wang, Zhengdong Xiao, Rui Xie, Longjiu Luo, Qianrong Wang, Gurong Cui, Honglin Qiao, Wenlian Lu
NLP Large Language Models Reinforcement Learning
  • Introduces Latent Logic Augmentation to enhance decision-making capabilities in LLMs.
  • Develops a Multiple Ground Truths dataset to reduce noise and capture semantic diversity.
  • Presents a Hybrid Reward mechanism for efficient reinforcement learning.
  • Demonstrates improved stability and performance in real-world Cloud service tasks.
Read more
GAPSL: A Gradient-Aligned Parallel Split Learning on Heterogeneous Data
Zheng Lin, Ons Aouedi, Wei Ni, Symeon Chatzinotas, Xianhao Chen
Federated Learning Efficient ML Optimization
  • GAPSL mitigates gradient directional inconsistency in parallel split learning.
  • The framework includes Leader Gradient Identification (LGI) and Gradient Direction Alignment (GDA) components.
  • GAPSL outperforms state-of-the-art methods in training accuracy and latency.
  • The approach is particularly beneficial for resource-constrained client devices in federated learning scenarios.
Read more
Neural Galerkin Normalizing Flow for Transition Probability Density Functions of Diffusion Models
Riccardo Saporiti, Fabio Nobile
Generative Models Theory Optimization
  • Introduces Neural Galerkin Normalizing Flow (NGNF) for approximating TPDFs of diffusion processes.
  • Ensures structural integrity of the solution through Normalizing Flows, maintaining positivity and mass conservation.
  • Derives a system of ODEs for time evolution of flow parameters, enhancing the learning of causal relationships.
  • Utilizes adaptive sampling to effectively evaluate Fokker-Planck residuals in high-dimensional PDEs.
Read more
Enactor: From Traffic Simulators to Surrogate World Models
Yash Ranjan, Rahul Sengupta, Anand Rangarajan, Sanjay Ranka
Generative Models Reinforcement Learning Robotics
  • Introduction of Enactor, a transformer-based generative model for traffic simulation.
  • Model captures complex actor-actor interactions and generates physically consistent trajectories.
  • Demonstrated effectiveness in a live simulation setting with significant performance improvements over traditional models.
  • Requires fewer training samples compared to conventional agent-centric approaches.
Read more
HISR: Hindsight Information Modulated Segmental Process Rewards For Multi-turn Agentic Reinforcement Learning
Zhicong Lu, Zichuan Lin, Wei Jia, Changyuan Tian, Deheng Ye, Peiguang Li, Li Jin, Nayu Liu, Guangluan Xu, Wei Feng
Reinforcement Learning Large Language Models
  • HISR improves credit assignment in multi-turn RL by aligning rewards with sub-goals.
  • A segment-level process reward model is introduced to avoid overly fine-grained reward allocation.
  • The hindsight model captures action importance based on trajectory outcomes.
  • Extensive experiments show HISR achieves state-of-the-art performance on agentic benchmarks.
Read more
A Family of Adaptive Activation Functions for Mitigating Failure Modes in Physics-Informed Neural Networks
Krishna Murari
Theory Optimization Efficient ML
  • Introduction of adaptive wavelet-based activation functions to improve PINNs.
  • Systematic evaluation across multiple PDE classes shows enhanced performance.
  • Demonstrated robustness and accuracy over traditional activation functions.
  • Validated against various models including PINNsFormer and other deep learning architectures.
Read more
HEP Statistical Inference for UAV Fault Detection: CLs, LRT, and SBI Applied to Blade Damage
Khushiyant
Robotics Time Series Theory
  • Introduces a unified inference framework for UAV fault detection using HEP statistical methods.
  • Achieves high detection rates and low false alarm rates through the application of LRT, CLs, and SNPE.
  • Demonstrates superior performance compared to traditional fault detection methods like CUSUM and autoencoders.
  • Provides calibrated uncertainty estimates for fault severity, enhancing decision-making for operators.
Read more
An Optimised Greedy-Weighted Ensemble Framework for Financial Loan Default Prediction
Ezekiel Nii Noye Nortey, Jones Asante-Koranteng, Marcellin Atemkeng, Theophilus Ansah-Narh, David Mensah, Rebecca Davis, Ravenhill Adjetey Laryea
Optimization Interpretability
  • Proposes a dynamic ensemble framework for loan default prediction that adapts to changing data conditions.
  • Utilizes Particle Swarm Optimization for hyperparameter tuning of multiple classifiers.
  • Achieves significant improvements in predictive performance over traditional models and static ensemble methods.
  • Identifies key features influencing loan defaults, enhancing interpretability of the model.
Read more
Taming Epilepsy: Mean Field Control of Whole-Brain Dynamics
Ming Li, Ting Gao, Jingqiao Duan
Graph Learning Optimization Theory
  • Introduction of the GK-MFG framework for controlling epileptic seizures.
  • Integration of Reservoir Computing and graph-theoretic modeling for enhanced robustness.
  • Use of graph Laplacian constraints to respect the brain's functional topology.
  • Demonstration of the framework's effectiveness in suppressing seizures in high-dimensional networks.
Read more
InfoMamba: An Attention-Free Hybrid Mamba-Transformer Model
Youjin Wang, Jiaqiao Zhao, Rong Fu, Run Zhou, Ruizhe Zhang, Jiani Liang, Suisuai Cao, Feng Zhou
Efficient ML Computer Vision NLP
  • Introduces InfoMamba, an attention-free hybrid architecture for sequence modeling.
  • Balances fine-grained local interactions with long-range dependencies without quadratic complexity.
  • Employs a concept-bottleneck linear filtering layer and information-maximizing fusion for dynamic context integration.
  • Demonstrates superior performance compared to existing Transformer and SSM models across multiple benchmarks.
Read more
LLM-Augmented Computational Phenotyping of Long Covid
Jing Wang, Jie Shen, Amar Sra, Qiaomin Xie, Jeremy C Weiss
NLP Large Language Models Time Series
  • Introduction of 'Grace Cycle', an LLM-augmented framework for phenotyping Long Covid.
  • Identification of three clinical phenotypes: Protected, Responder, and Refractory.
  • Significant differences in symptom severity and treatment response among identified phenotypes.
  • Demonstration of LLMs' capability to enhance clinical data analysis and hypothesis testing.
Read more