"Learning Semantic Latent Directions for Accurate and Controllable Human Motion Prediction" (ECCV 2024)
This repo contains the official implementation of the paper:
Learning Semantic Latent Directions for Accurate and Controllable Human Motion Prediction
ECCV 2024 [arxiv]
- Python >= 3.8
- PyTorch >= 1.9
- Tensorboard
- matplotlib
- tqdm
- argparse
We adapt the data preprocessing from GSPS.
- We follow the data preprocessing steps (DATASETS.md) inside the VideoPose3D repo.
- Given the processed dataset, we further compute the multi-modal future for each motion sequence. All data needed can be downloaded from Google Drive and place all the dataset in
data
folder inside the root of this repo.
- All pretrain models can be downloaded from Google Drive and place all the pretrain models in
results
folder inside the root of this repo.
We have used the following commands for training the network on Human3.6M or HumanEva-I with skeleton representation:
python train_nf.py --cfg [h36m/humaneva] --gpu_index 0
python main.py --cfg [h36m/humaneva] --gpu_index 0
To test on the pretrained model, we have used the following commands:
python main.py --cfg [h36m/humaneva] --mode test --iter 500 --gpu_index 0
For visualizing from a pretrained model, we have used the following commands:
python main.py --cfg [h36m/humaneva] --mode viz --iter 500 --gpu_index 0
This code is based on the implementations of STARS.
If you find this work useful in your research, please cite:
@article{xu2024learning,
title={Learning Semantic Latent Directions for Accurate and Controllable Human Motion Prediction},
author={Xu, Guowei and Tao, Jiale and Li, Wen and Duan, Lixin},
journal={arXiv preprint arXiv:2407.11494},
year={2024}
}
This repo is distributed under an MIT LICENSE