xxlong0 / Wonder3D
- воскресенье, 29 октября 2023 г. в 00:00:01
A cross-domain diffusion model for 3D reconstruction from a single image
Single Image to 3D using Cross-Domain Diffusion
Wonder3D reconstructs highly-detailed textured meshes from a single-view image in only 2 ∼ 3 minutes. Wonder3D first generates consistent multi-view normal maps with corresponding color images via a cross-domain diffusion model, and then leverages a novel normal fusion method to achieve fast and high-quality reconstruction.
requirements.txt
.conda create -n wonder3d
conda activate wonder3d
pip install -r requirements.txt
Install tiny-cuda-nn PyTorch extension for mesh extraction: pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch
Wonder3D
|-- ckpts
|-- unet
|-- scheduler.bin
...
rembg
to remove the backgrounds.# !pip install rembg
import rembg
result = rembg.remove(result)
result.show()
./outputs
. (we use rembg to remove backgrounds of the results, but the segmemtations are not always perfect.)accelerate launch --config_file 1gpu.yaml test_mvdiffusion_seq.py \
--config mvdiffusion-joint-ortho-6views.yaml
or
bash run_test.sh
cd ./instant-nsr-pl
bash run.sh output_folder_path scene_name
Our generated normals and color images are defined in orthographic views, so the reconstructed mesh is also in orthographic camera space. If you use MeshLab to view the meshes, you can click Toggle Orthographic Camera
in View
tab.
If you find this repository useful in your project, please cite the following work. :)
@misc{long2023wonder3d,
title={Wonder3D: Single Image to 3D using Cross-Domain Diffusion},
author={Xiaoxiao Long and Yuan-Chen Guo and Cheng Lin and Yuan Liu and Zhiyang Dou and Lingjie Liu and Yuexin Ma and Song-Hai Zhang and Marc Habermann and Christian Theobalt and Wenping Wang},
year={2023},
eprint={2310.15008},
archivePrefix={arXiv},
primaryClass={cs.CV}
}