This is the official implementation of MVGamba, a novel Multi-View Gamba Model designed for unified 3D content generation.
MVGamba_video_demo.mp4
Arxiv | [Weights]
- Release MVGamba-3DGS training and inference code.
- Release MVGamba-2DGS training and inference code, please refer to the surfel branch.
- Release pretrained checkpoints.
# xformers is required! please refer to https://github.com/facebookresearch/xformers for details.
# for example, we use torch 2.1.0 + cuda 11.8
pip install torch==2.1.0 torchvision==0.16.0 torchaudio==2.1.0 --index-url https://download.pytorch.org/whl/cu118
pip install causal-conv1d==1.2.0 mamba-ssm
git clone --recursive [email protected]:SkyworkAI/MVGamba.git
# a modified 3D gaussian splatting (+ depth, alpha rendering)
pip install ./submodules/diff-gaussian-rasterization
# 2D gaussian surfel
pip install ./submodules/diff-surfel-rasterization
# for mesh extraction
pip install git+https://github.com/NVlabs/nvdiffrast
# other dependencies
pip install -r requirements.txt
NOTE: Due to company property restrictions, we are unable to provide the full training data used for training MVGamba. Here, we alternatively follow the practice by @kiui(https://github.com/ashawkey) and provide the ~80K subset of Objaverse and in objaverse_filter. please check and modify the dataset implementation!
# training (use single node training)
accelerate launch --config_file acc_configs/gpu8.yaml main.py mvgamba --workspace /root/Results/workspace_mvgamba
# training (use multi-node training)
bash ./scripts/mvgamba_dist.sh
For MVDream and ImageDream, we use a diffusers implementation. The required model weights are downloaded automatically.
bash ./scripts/mvgamba_infer.sh
This work is built on many amazing research works and open-source projects, thanks a lot to all the authors for sharing!
@article{yi2024mvgamba,
title={MVGamba: Unify 3D Content Generation as State Space Sequence Modeling},
author={Yi, Xuanyu and Wu, Zike and Shen, Qiuhong and Xu, Qingshan and Zhou, Pan and Lim, Joo-Hwee and Yan, Shuicheng and Wang, Xinchao and Zhang, Hanwang},
journal={arXiv preprint arXiv:2406.06367},
year={2024}
}
Please also check our another project for ultra fast single view 3D generation Gamba. The code and pretrained weights have been released in https://github.com/SkyworkAI/Gamba.
@article{shen2024gamba,
title={Gamba: Marry gaussian splatting with mamba for single view 3d reconstruction},
author={Shen, Qiuhong and Wu, Zike and Yi, Xuanyu and Zhou, Pan and Zhang, Hanwang and Yan, Shuicheng and Wang, Xinchao},
journal={arXiv preprint arXiv:2403.18795},
year={2024}
}