The paper presents the HairFast method, a new approach for transferring hairstyles from a reference image to an input photo. The key challenges in this task include adapting to various photo poses, the sensitivity of hairstyles, and the lack of objective metrics.
The paper first discusses existing approaches, which can be divided into optimization-based and encoder-based methods. Optimization-based methods have good quality but are slow, while encoder-based methods are fast but suffer from poor quality and low resolution.
The HairFast method uniquely solves these problems by using a new architecture with four modules: embedding, alignment, blending, and post-processing. The embedding module obtains various latent representations of the input images, including in the FS and W+ spaces of StyleGAN. The alignment module transfers the desired hairstyle shape, the blending module transfers the desired hair color, and the post-processing module restores details lost during the embedding step.
Extensive experiments on the CelebA-HQ dataset show that HairFast achieves comparable or better results than state-of-the-art optimization-based methods in terms of realism metrics like FID and FIDCLIP, while having inference times comparable to the fastest encoder-based method, HairCLIP. The method also handles large pose differences better than previous approaches.
The paper concludes by discussing the limitations of the current method and opportunities for future work, such as enabling more flexible hairstyle editing capabilities.
다른 언어로
소스 콘텐츠 기반
arxiv.org
더 깊은 질문