This research paper introduces MVLight, a novel light-conditioned multi-view diffusion model for text-to-3D generation. The authors address the challenge of decoupling light-independent and lighting-dependent components in 3D models to enhance their quality and relighting performance.
The study aims to develop a method for generating high-quality, relightable 3D models from textual descriptions by incorporating lighting conditions directly into the generation process.
The researchers propose MVLight, a multi-view diffusion model that integrates lighting information through HDR images. They decouple HDR images into high-frequency and low-frequency components, embedding them into the model through a light cross-attention module. MVLight generates multi-view consistent images, albedo, and normal maps under specified lighting conditions. The model is trained on a custom dataset (XMV L) of objects captured from multiple viewpoints under various lighting conditions, along with textual descriptions. The researchers utilize Score Distillation Sampling (SDS) with a two-stage optimization process: first synthesizing geometry and appearance, then fine-tuning PBR materials for enhanced relighting.
MVLight significantly advances text-to-3D generation by enabling the creation of high-quality, relightable 3D models with improved geometric accuracy and superior relighting capabilities. The direct integration of lighting conditions into the generation process through a light-conditioned multi-view diffusion model proves to be an effective approach for enhancing 3D model synthesis.
This research contributes to the field of computer graphics by introducing a novel approach for generating relightable 3D models from text. MVLight has the potential to impact various industries, including gaming, virtual reality, and animation, by simplifying the creation of realistic and customizable 3D assets.
While MVLight successfully generates multi-view consistent outputs for different modalities, ensuring alignment between these modalities remains a challenge. Future research could explore methods for improving modality alignment without compromising output quality. Additionally, investigating the generalization capabilities of MVLight to unseen object categories and more complex lighting scenarios could further enhance its applicability.
Іншою мовою
із вихідного контенту
arxiv.org
Ключові висновки, отримані з
by Dongseok Shi... о arxiv.org 11-19-2024
https://arxiv.org/pdf/2411.11475.pdfГлибші Запити