The paper presents a novel method called HAHA (Highly Articulated Gaussian Human Avatars with Textured Mesh Prior) for generating animatable human avatars from monocular input videos. The key idea is to learn a joint representation using Gaussian splatting and textured mesh, where the textured mesh is used to represent the body surface and Gaussians are used to capture out-of-mesh details like hair and clothing.
The method consists of three stages:
The authors demonstrate that HAHA can achieve reconstruction quality on par with state-of-the-art methods on the SnapshotPeople dataset while using significantly fewer Gaussians (up to 3 times fewer). They also show that HAHA outperforms previous methods on the more challenging X-Humans dataset, both quantitatively and qualitatively, especially in handling highly articulated body parts like fingers.
The key contributions of the work are:
Para outro idioma
do conteúdo fonte
arxiv.org
Principais Insights Extraídos De
by David Svitov... às arxiv.org 04-02-2024
https://arxiv.org/pdf/2404.01053.pdfPerguntas Mais Profundas