ROMP is a one-stage method for monocular multi-person 3D mesh recovery in real time. | BEV further explores multi-person depth relationships and supports all age groups. |
[Paper] [Video] | [Project Page] [Paper] [Video] [RH Dataset] |
![]() |
![]() |
We provide cross-platform API (installed via pip) to run ROMP & BEV on Linux / Windows / Mac.
2022/06/21: Training & evaluation code of BEV is released. Please update the model_data.
2022/05/16: simple-romp v1.0 is released to support tracking, calling in python, exporting bvh, and etc.
2022/04/14: Inference code of BEV has been released in simple-romp v0.1.0.
2022/04/10: Adding onnx support, with faster inference speed on CPU/GPU.
Old logs
Please use simple-romp for inference, the rest code is just for training.
pip install --upgrade setuptools numpy==1.22 cython
pip install --upgrade simple-romp
For more details, please refer to install.md.
Please refer to this guidance for inference & export (fbx/glb/bvh).
For training, please refer to installation.md for full installation. Please prepare the training datasets following dataset.md, and then refer to train.md for training.
Please refer to romp_evaluation.md and bev_evaluation.md for evaluation on benchmarks.
[Blender addon]: Yan Chuanhang created a Blender-addon to drive a 3D character in Blender using ROMP from image, video or webcam input.
[VMC protocol]: Vivien Richter implemented a VMC (Virtual Motion Capture) protocol support for different Motion Capture solutions with ROMP.
Please refer to docker.md
Welcome to submit issues for the bugs.
This repository is currently maintained by Yu Sun.
We thank Peng Cheng for his constructive comments on Center map training.
ROMP has also benefited from many developers, including
- Marco Musy : help in the textured SMPL visualization.
- Gavin Gray : adding support for an elegant context manager to run code in a notebook.
- VLT Media and Vivien Richter : adding support for running on Windows & batch_videos.py.
- Chuanhang Yan : developing an addon for driving character in Blender.
- Tian Jin: help in simplified smpl and fast rendering (realrender).
- ZhengdiYu : helpful discussion on optimizing the implementation details.
- Ali Yaghoubian : add Docker file for simple-romp.
@InProceedings{BEV,
author = {Sun, Yu and Liu, Wu and Bao, Qian and Fu, Yili and Mei, Tao and Black, Michael J},
title = {Putting People in their Place: Monocular Regression of 3D People in Depth},
booktitle = {CVPR},
year = {2022}}
@InProceedings{ROMP,
author = {Sun, Yu and Bao, Qian and Liu, Wu and Fu, Yili and Michael J., Black and Mei, Tao},
title = {Monocular, One-stage, Regression of Multiple 3D People},
booktitle = {ICCV},
year = {2021}}
We thank all contributors for their help!
This work was supported by the National Key R&D Program of China under Grand No. 2020AAA0103800.
Disclosure: MJB has received research funds from Adobe, Intel, Nvidia, Facebook, and Amazon and has financial interests in Amazon, Datagen Technologies, and Meshcapade GmbH. While he was part-time at Amazon during this project, his research was performed solely at Max Planck.