Parallel Simulation for Log-concave Sampling and Score-based Diffusion Models

Published: 01 May 2025, Last Modified: 18 Jun 2025ICML 2025 spotlightposterEveryoneRevisionsBibTeXCC BY 4.0
TL;DR: We propose new parallel algorithms for log-concave sampling and diffusion models, and rigorously prove that our algorithms enjoy $\log d$ iteration complexity.
Abstract: Sampling from high-dimensional probability distributions is fundamental in machine learning and statistics. As datasets grow larger, computational efficiency becomes increasingly important, particularly in reducing *adaptive complexity*, namely the number of sequential rounds required for sampling algorithms. While recent works have introduced several parallelizable techniques, they often exhibit suboptimal convergence rates and remain significantly weaker than the latest lower bounds for log-concave sampling. To address this, we propose a novel parallel sampling method that improves adaptive complexity dependence on dimension $d$ reducing it from $\widetilde{\mathcal{O}}(\log^2 d)$ to $\widetilde{\mathcal{O}}(\log d)$. Our approach builds on parallel simulation techniques from scientific computing.
Lay Summary: Sampling from high-dimensional probability distributions is a fundamental task in machine learning and statistics, essential for applications like Bayesian inference and generative modeling. However, existing sampling methods often require numerous sequential steps, limiting their efficiency and scalability, especially as datasets grow larger. Our research introduces a novel parallel sampling algorithm that significantly reduces the number of sequential steps—known as adaptive complexity—required to generate high-quality samples. By leveraging techniques from scientific computing, our method improves the adaptive complexity from approximately the square of the logarithm of the data's dimensionality to just the logarithm. As a result, our method speeds up sampling for both log-concave distributions and diffusion models. This advancement enables more efficient processing of large-scale data, facilitating faster and more scalable applications in machine learning tasks such as generative modeling and Bayesian inference.
Primary Area: Probabilistic Methods->Monte Carlo and Sampling Methods
Keywords: parallel sampling, log-concave sampling, diffusion model
Submission Number: 9486
Loading