Conceptos Básicos
A novel approach for generating photo-realistic and animatable human avatars from monocular input videos by learning a joint representation using Gaussian splatting and textured mesh.
Resumen
The paper presents a novel method called HAHA (Highly Articulated Gaussian Human Avatars with Textured Mesh Prior) for generating animatable human avatars from monocular input videos. The key idea is to learn a joint representation using Gaussian splatting and textured mesh, where the textured mesh is used to represent the body surface and Gaussians are used to capture out-of-mesh details like hair and clothing.
The method consists of three stages:
- In the first stage, a full Gaussian representation of the avatar is learned by optimizing the Gaussian parameters and fine-tuning the SMPL-X pose and shape.
- In the second stage, a textured mesh representation of the avatar is learned by optimizing the texture while keeping the SMPL-X parameters fixed.
- In the final stage, the Gaussian and textured mesh representations are merged, and an unsupervised method is used to remove unnecessary Gaussians by learning their opacity.
The authors demonstrate that HAHA can achieve reconstruction quality on par with state-of-the-art methods on the SnapshotPeople dataset while using significantly fewer Gaussians (up to 3 times fewer). They also show that HAHA outperforms previous methods on the more challenging X-Humans dataset, both quantitatively and qualitatively, especially in handling highly articulated body parts like fingers.
The key contributions of the work are:
- The use of a joint representation with Gaussians and textured mesh to increase the efficiency of rendering human avatars.
- An unsupervised method for significantly reducing the number of Gaussians in the scene through the use of a textured mesh.
- The ability to efficiently handle the animation of highly articulated body parts like hands without any additional engineering.
Estadísticas
The paper does not contain any key metrics or important figures to support the author's key logics.
Citas
The paper does not contain any striking quotes supporting the author's key logics.