Face personalization involves inserting specific faces into text-to-image models, but previous methods struggle with overfitting. Face2Diffusion introduces innovative components like multi-scale identity encoder, expression guidance, and class-guided denoising regularization to address these challenges. The method aims to balance identity similarity and editability in generated images, as demonstrated through experiments on the FaceForensics++ dataset. By removing identity-irrelevant information from training data, Face2Diffusion significantly improves the trade-off between identity- and text-fidelity compared to existing methods.
他の言語に翻訳
原文コンテンツから
arxiv.org
深掘り質問