Core Concepts
PLANNER combines latent semantic diffusion with autoregressive generation to create fluent text with global control over paragraphs.
Abstract
Autoregressive models can lead to repetitive and low-quality output due to exposure bias.
PLANNER proposes a model that combines latent semantic diffusion with autoregressive generation for better text generation.
The model uses a planning module for semantic paragraph embeddings and a decoding module for generating text.
PLANNER is evaluated on various conditional generation tasks, showing effectiveness in generating high-quality long-form text efficiently.
Stats
Autoregressive models trained with teacher forcing strategy are considered the gold standard for text generation.
Diffusion models provide an alternative solution by revisiting and revising output iteratively.
Quotes
"The model achieves this by combining an autoregressive “decoding” module with a “planning” module that uses latent diffusion to generate semantic paragraph embeddings in a coarse-to-fine manner."