The paper aims to establish a full generation error analysis for diffusion-based generative models by considering both the training and sampling processes.
For the training process, the authors focus on the denoising score matching objective and prove the exponential convergence of its gradient descent training dynamics. They develop a new method to establish a key lower bound of the gradient under the semi-smoothness framework.
For the sampling process, the authors extend the existing sampling error analysis to the variance exploding setting, only requiring the data distribution to have finite second moment. Their result applies to various time and variance schedules, and implies a sharp almost linear complexity in terms of data dimension under the optimal time schedule.
By combining the training and sampling analyses, the authors conduct a full error analysis of diffusion models. They qualitatively derive the theory for choosing the noise distribution and weighting in the training objective, which coincides with previous empirical findings. Additionally, they develop a theory for choosing time and variance schedules based on both training and sampling, showing that the optimal schedule depends on whether the score error or the sampling error dominates.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Yuqing Wang,... at arxiv.org 09-10-2024
https://arxiv.org/pdf/2406.12839.pdfDeeper Inquiries