Skip to content

ptrvilya/object-popup

Repository files navigation

Object pop-up: Can we infer 3D objects and their poses from human interactions alone?

Ilya A. PetrovRiccardo MarinJulian ChibaneGerard Pons-Moll

CVPR 2023

Project Teaser

Paper PDF Project Page YouTube video

Environment

The code was tested under Ubuntu 22.04, Python 3.10, CUDA 11.8, PyTorch 2.0.1.
Use the following command to create a conda environment with necessary dependencies:

conda env create -f environment.yml

Data downloading and processing

The steps are described in docs/data.md.

Pre-trained models and evaluation

Pre-trained models can be obtained from the link. With the commands:

wget https://nc.mlcloud.uni-tuebingen.de/index.php/s/PG8wZ5HRKytEY8S/download/object_pop_up_noclass.tar -P ./assets 
wget https://nc.mlcloud.uni-tuebingen.de/index.php/s/Dfx9rfQ2tW4ZsEY/download/object_pop_up_class.tar -P ./assets

Use the following commands to run evaluation:

# model without class prediction (assumes 24GB GPU memory)
python evaluate.py scenarios/gb_PNv2_noclass.toml -b 64 -w 20 -d grab behave -g -rc ./assets/object_pop_up_noclass.pth -c configs/smplh.toml
# model with class prediction (assumes 24GB GPU memory)
python evaluate.py scenarios/gb_PNv2_class.toml -b 64 -w 20 -d grab behave -g -rc ./assets/object_pop_up_class.pth -c configs/smplh.toml

Training

Use the following command to run the training:

python train.py scenarios/gb_PNv2_noclass.toml -b 32 -w 10 -nowb -ep 0001_smplh -c configs/smplh.toml

Citation

@inproceedings{petrov2023popup,
   title={Object pop-up: Can we infer 3D objects and their poses from human interactions alone?},
   author={Petrov, Ilya A and Marin, Riccardo and Chibane, Julian and Pons-Moll, Gerard},
   booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
   year={2023}
}

Acknowledgements

This project benefited from the following resources: