Core Concepts
VOLTAは、TransformerとVAEを組み合わせて生成的多様性を向上させる革新的なフレームワークです。
Abstract
Abstract:
Transformer models have led to success in natural language generation.
VOLTA framework bridges Transformer with VAE to enhance generative diversity.
Incorporates InfoGAN-style latent codes for input-independent variability.
Introduction:
Transformer models prioritize quality over diversity in autoregressive text generation.
Generative diversity is crucial in NLG, distinct from paraphrasing.
Context:
Early attempts like diverse beam search enhanced diversity but had limitations.
VAE framework addresses low-diversity issue by encoding inputs into lower-dimensional latent variables.
Data Extraction:
Latent Space Variable Code N(μ, σ²)
Optimus adopts BERT as the VAE encoder and GPT-2 as the VAE decoder.
Quotations:
"Generative diversity is distinct from mere paraphrasing, as it encompasses not only altered syntax but also varied semantics."
Stats
Latent Space Variable Code N(μ, σ²)
Quotes
"Generative diversity is distinct from mere paraphrasing, as it encompasses not only altered syntax but also varied semantics."