核心概念
The proposed observation-guided diffusion probabilistic model (OGDM) effectively addresses the trade-off between quality control and fast sampling in diffusion-based image generation by integrating the guidance of the observation process with the Markov chain in a principled way.
摘要
The paper presents a novel diffusion-based image generation method called the observation-guided diffusion probabilistic model (OGDM). The key idea is to reestablish the training objective by integrating the guidance of the observation process with the Markov chain in a principled way. This is achieved by introducing an additional loss term derived from the observation based on a conditional discriminator on noise level, which employs a Bernoulli distribution indicating whether its input lies on the (noisy) real manifold or not.
The authors show that this strategy allows them to optimize the more accurate negative log-likelihood induced in the inference stage, especially when the number of function evaluations is limited. The proposed training scheme is also advantageous even when incorporated only into the fine-tuning process, and it is compatible with various fast inference strategies since their method yields better denoising networks using the exactly the same inference procedure without incurring extra computational cost.
The authors demonstrate the effectiveness of their training algorithm using diverse inference techniques on strong diffusion model baselines. The results show that the proposed approach outperforms the baseline models in terms of both FID and recall scores, especially when the number of function evaluations (NFEs) is small.
統計資料
The paper does not contain any key metrics or important figures to support the author's key logics.
引述
The paper does not contain any striking quotes supporting the author's key logics.