MVDream: Multi-view Diffusion for 3D Generation
Yichun Shi, Peng Wang, Jianglong Ye, Mai Long, Kejie Li, Xiao Yang
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/bytedance/mvdreampytorch★ 978
- github.com/bytedance/MVDream-threestudiojax★ 550
- github.com/nirvanalan/ln3diffpytorch★ 227
- github.com/MindCode-4/code-10/tree/main/MVDmindspore★ 0
Abstract
We introduce MVDream, a diffusion model that is able to generate consistent multi-view images from a given text prompt. Learning from both 2D and 3D data, a multi-view diffusion model can achieve the generalizability of 2D diffusion models and the consistency of 3D renderings. We demonstrate that such a multi-view diffusion model is implicitly a generalizable 3D prior agnostic to 3D representations. It can be applied to 3D generation via Score Distillation Sampling, significantly enhancing the consistency and stability of existing 2D-lifting methods. It can also learn new concepts from a few 2D examples, akin to DreamBooth, but for 3D generation.