Reverse Transition Kernel: A Flexible Framework to Accelerate Diffusion Inference
CoRR(2024)
Abstract
To generate data from trained diffusion models, most inference algorithms,
such as DDPM, DDIM, and other variants, rely on discretizing the reverse SDEs
or their equivalent ODEs. In this paper, we view such approaches as decomposing
the entire denoising diffusion process into several segments, each
corresponding to a reverse transition kernel (RTK) sampling subproblem.
Specifically, DDPM uses a Gaussian approximation for the RTK, resulting in low
per-subproblem complexity but requiring a large number of segments (i.e.,
subproblems), which is conjectured to be inefficient. To address this, we
develop a general RTK framework that enables a more balanced subproblem
decomposition, resulting in Õ(1) subproblems, each with strongly
log-concave targets. We then propose leveraging two fast sampling algorithms,
the Metropolis-Adjusted Langevin Algorithm (MALA) and Underdamped Langevin
Dynamics (ULD), for solving these strongly log-concave subproblems. This gives
rise to the RTK-MALA and RTK-ULD algorithms for diffusion inference. In theory,
we further develop the convergence guarantees for RTK-MALA and RTK-ULD in total
variation (TV) distance: RTK-ULD can achieve ϵ target error within
𝒪̃(d^1/2ϵ^-1) under mild conditions, and RTK-MALA
enjoys a 𝒪(d^2log(d/ϵ)) convergence rate under slightly
stricter conditions. These theoretical results surpass the state-of-the-art
convergence rates for diffusion inference and are well supported by numerical
experiments.
MoreTranslated text
AI Read Science
Must-Reading Tree
Example
![](https://originalfileserver.aminer.cn/sys/aminer/pubs/mrt_preview.jpeg)
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined