The paper presents a novel method called HAHA (Highly Articulated Gaussian Human Avatars with Textured Mesh Prior) for generating animatable human avatars from monocular input videos. The key idea is to learn a joint representation using Gaussian splatting and textured mesh, where the textured mesh is used to represent the body surface and Gaussians are used to capture out-of-mesh details like hair and clothing.
The method consists of three stages:
The authors demonstrate that HAHA can achieve reconstruction quality on par with state-of-the-art methods on the SnapshotPeople dataset while using significantly fewer Gaussians (up to 3 times fewer). They also show that HAHA outperforms previous methods on the more challenging X-Humans dataset, both quantitatively and qualitatively, especially in handling highly articulated body parts like fingers.
The key contributions of the work are:
Naar een andere taal
vanuit de broninhoud
arxiv.org
Belangrijkste Inzichten Gedestilleerd Uit
by David Svitov... om arxiv.org 04-02-2024
https://arxiv.org/pdf/2404.01053.pdfDiepere vragen