The paper introduces Hash3D, a novel approach to accelerate diffusion-based 3D generation models without any additional training. The key insight is that feature maps rendered from nearby camera positions and diffusion time-steps exhibit a high degree of redundancy.
Hash3D employs an adaptive grid-based hashing mechanism to efficiently store and retrieve these similar features, significantly reducing the number of redundant calculations required during the optimization process. This feature-sharing strategy not only speeds up the 3D generation but also enhances the smoothness and view consistency of the synthesized 3D objects.
The authors extensively evaluate Hash3D by integrating it with a diverse range of text-to-3D and image-to-3D models. The results demonstrate that Hash3D can accelerate the optimization process by 1.3 to 4 times without compromising performance. Additionally, the integration of Hash3D with 3D Gaussian Splatting leads to a substantial reduction in processing time, cutting down text-to-3D to about 10 minutes and image-to-3D to roughly 30 seconds.
The paper highlights the following key contributions:
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Xingyi Yang,... at arxiv.org 04-10-2024
https://arxiv.org/pdf/2404.06091.pdfDeeper Inquiries