Jason Y. Zhang*, Sam Pepose*, Hanbyul Joo, Deva Ramanan, Jitendra Malik, and Angjoo Kanazawa.
[arXiv
]
[Project Page
]
[Bibtex
]
In ECCV 2020
- Python (tested on 3.7)
- Pytorch (tested on 1.4)
- Neural Mesh Renderer
- PointRend (from Detectron)
- BodyMocap
There is currently no CPU-only support.
Our code is released under CC BY-NC 4.0. However, our code depends on other libraries, including SMPL, which each have their own respective licenses that must also be followed.
We recommend using a conda environment:
conda create -n phosa python=3.7
conda activate phosa
Install the torch version that corresponds to your version of CUDA, eg for CUDA 10.0, use:
conda install pytorch=1.4.0 torchvision=0.5.0 cudatoolkit=10.0 -c pytorch
Alternatively, you can check out our interactive Colab Notebook.
Install the fast version of Neural Mesh Renderer:
mkdir -p external
git clone https://github.com/JiangWenPL/multiperson.git external/multiperson
pip install external/multiperson/neural_renderer
Install Detectron2:
mkdir -p external
git clone https://github.com/facebookresearch/detectron2.git external/detectron2
pip install external/detectron2
Install FrankMocap (The body module is the same regressor trained on EFT data that we used in the paper):
mkdir -p external
git clone https://github.com/facebookresearch/frankmocap.git external/frankmocap
sh scripts/download_data_body_module.sh
You will also need to download the SMPL model from the SMPL website. Make an account, and download the neutral model basicModel_neutral_lbs_10_207_0_v1.0.0.pkl and place it in extra_data/smpl/basicModel_neutral_lbs_10_207_0_v1.0.0.pkl
If you did not clone detectron2 and frankmocap in the external
directory, you will need to update the paths in the constants file.
Currently, the mesh interpenetration loss is not included, so the results may look slightly different from the paper.
The repository only includes a bicycle mesh that we created. For other object categories and mesh instances, you will need to download your own meshes. We list some recommended sources here.
python -m demo.py --filename input/000000038829.jpg
We also have a Colab Notebook to interactively visualize the outputs.
If you use find this code helpful, please consider citing:
@InProceedings{zhang2020phosa,
title = {Perceiving 3D Human-Object Spatial Arrangements from a Single Image in the Wild},
author = {Zhang, Jason Y. and Pepose, Sam and Joo, Hanbyul and Ramanan, Deva and Malik, Jitendra and Kanazawa, Angjoo},
booktitle = {European Conference on Computer Vision (ECCV)},
year = {2020},
}