The paper investigates the stability of iteratively retraining generative models on a mix of real data and synthetic data generated by the current model. It proposes a theoretical framework to study this setting and proves the following key insights:
Under the condition that the initial generative model is sufficiently well-trained and the proportion of real data is large enough, the iterative retraining procedure is stable and the model converges to the optimal generative model (Theorem 1).
If the above conditions are not met, the iterative retraining can lead to model collapse, where the generative model degrades to outputting a single point (Proposition 1).
With finite sampling, the paper shows that the iteratively retrained model remains within a neighborhood of the optimal generative model, with the error decomposing into optimization error, statistical error, and the iterative retraining error (Theorem 2).
The paper validates the theoretical findings through experiments on synthetic datasets as well as natural image datasets like CIFAR-10 and FFHQ, using powerful generative models like normalizing flows and diffusion models.
In eine andere Sprache
aus dem Quellinhalt
arxiv.org
Wichtige Erkenntnisse aus
by Quentin Bert... um arxiv.org 04-03-2024
https://arxiv.org/pdf/2310.00429.pdfTiefere Fragen