DreamFusion: Text-to-3D using 2D Diffusion

Ben Poole
Google Research
Ajay Jain
UC Berkeley
Jonathan T. Barron
Google Research
Ben Mildenhall
Google Research
Paper Project Gallery

Abstract

Recent breakthroughs in text-to-image synthesis have been driven by diffusion models trained on billions of image-text pairs. Adapting this approach to 3D synthesis would require large-scale datasets of labeled 3D assets and efficient architectures for denoising 3D data, neither of which currently exist. In this work, we circumvent these limitations by using a pretrained 2D text-to-image diffusion model to perform text-to-3D synthesis. We introduce a loss based on probability density distillation that enables the use of a 2D diffusion model as a prior for optimization of a parametric image generator. Using this loss in a DeepDream-like procedure, we optimize a randomly-initialized 3D model (a Neural Radiance Field, or NeRF) via gradient descent such that its 2D renderings from random angles achieve a low loss. The resulting 3D model of the given text can be viewed from any angle, relit by arbitrary illumination, or composited into any 3D environment. Our approach requires no 3D training data and no modifications to the image diffusion model, demonstrating the effectiveness of pretrained image diffusion models as priors.

Given a caption, DreamFusion generates relightable 3D objects with high-fidelity appearance, depth, and normals. Objects are represented as a Neural Radiance Field and leverage a pretrained text-to-image diffusion prior such as Imagen.

Generate 3D from text yourself!


Example generated objects

DreamFusion generates objects and scenes from diverse captions. Search through hundreds of generated assets in our full gallery.


Composing objects into a scene


Mesh exports

Our generated NeRF models can be exported to meshes using the marching cubes algorithm for easy integration into 3D renderers or modeling software.


How does DreamFusion work?

Given a caption, DreamFusion uses a text-to-image generative model called Imagen to optimize a 3D scene. We propose Score Distillation Sampling (SDS), a way to generate samples from a diffusion model by optimizing a loss function. SDS allows us to optimize samples in an arbitrary parameter space, such as a 3D space, as long as we can map back to images differentiably. We use a 3D scene parameterization similar to Neural Radiance Fields, or NeRFs, to define this differentiable mapping. SDS alone produces reasonable scene appearance, but DreamFusion adds additional regularizers and optimization strategies to improve geometry. The resulting trained NeRFs are coherent, with high-quality normals, surface geometry and depth, and are relightable with a Lambertian shading model.


Citation

@article{poole2022dreamfusion,
  author = {Poole, Ben and Jain, Ajay and Barron, Jonathan T. and Mildenhall, Ben},
  title = {DreamFusion: Text-to-3D using 2D Diffusion},
  journal = {arXiv},
  year = {2022},
}
主站蜘蛛池模板: 精品美女在线观看| 欧美xxxx三人交性视频| 亚洲欧美日韩中文字幕在线 | 国产国产人免费人成免费视频 | 日本人强jizz多人| 亚洲视频欧美视频| 国内精品久久久久精品| 欧美色图在线视频| 色一情一乱一伦一视频免费看| 久久成人国产精品| 国产成人综合美国十次| 天堂mv在线免费看| 波多野结衣忆青春| 1000部拍拍拍18免费网站| 久久婷婷人人澡人人喊人人爽 | 国产手机精品一区二区| 成年女人色毛片免费看| 精品乱码一区二区三区在线| JAPANESEHD熟女熟妇伦| 人妻体内射精一区二区| 国产精品福利影院| 日本免费人成视频在线观看| 澳门码资料2020年276期| 狠狠综合久久久久尤物丿| 漂亮人妻被黑人久久精品| 杨晨晨白丝mm131| 国产xxxx做受视频| swag合集120部| 日本黄色片下载| 国产91在线|日韩| 91色视频网站| 日本黄色免费观看| 亚洲国产成人久久| 男女边吃奶边做边爱视频 | 亚洲欧美日韩第一页| 精品国产一区二区三区久久狼 | 成人精品一区二区三区中文字幕| 日韩一区二区三区精品| 日韩欧美国产亚洲| 波多野结衣无内裤护士| 爆乳美女脱内衣18禁裸露网站|