Skip to content

Deep Multitask Architecture for Integrated 2D and 3D Human Sensing (CVPR 2017)

License

Notifications You must be signed in to change notification settings

alinionutpopa/dmhs

Repository files navigation

Deep Multitask Architecture for Integrated 2D and 3D Human Sensing

This package contains code for the Deep Multitask Human Sensing (DMHS) method, published in the CVPR 2017 paper Deep Multitask Architecture for Integrated 2D and 3D Human Sensing.

architecture

By using the software, you are agreeing to the terms of the license agreement.

Our software is built on top of the Caffe deep learning library used by the Convolutional Pose Machines method. The current version was developed by:

Alin-Ionut Popa , Mihai Zanfir and Cristian Sminchisescu

We provide a deep multitask architecture for fully automatic 2d and 3d human sensing (DMHS), including recognition and reconstruction, in monocular images. The system computes the figure-ground segmentation, semantically identifies the human body parts at pixel level, and estimates the 2d and 3d pose of the person. This software allows you to test our algorithm on your own images.

sample

If you use this code/model for your research, please cite the following paper:

@inproceedings{dmhs_cvpr17,
    author = {Alin-Ionut Popa and Mihai Zanfir and Cristian Sminchisescu},
    title  = {Deep Multitask Architecture for Integrated 2D and 3D Human Sensing},
    booktitle = {IEEE International Conference on Computer Vision and Pattern Recognition},
    year   = {2017}
}

Installation Guide

First, clone the project by running:

git clone --recursive https://github.com/alinionutpopa/dmhs.git

You need to compile the modified Caffe library in this repository. Instructions for Ubuntu 14.04 are included below. You can also consult the generic Caffe installation guide for further help.

1.1 Install dependencies

General dependencies
sudo apt-get install libprotobuf-dev libleveldb-dev libsnappy-dev libopencv-dev libhdf5-serial-dev protobuf-compiler
sudo apt-get install --no-install-recommends libboost-all-dev
CUDA (optional - needed only if you are planning to use a GPU for faster processing)

Install the correct CUDA driver and its SDK. Download CUDA SDK from Nvidia website.

You might need to blacklist some modules so that they do not interfere with the driver installation. You also need to uninstall your default Nvidia Driver first.

sudo apt-get install freeglut3-dev build-essential libx11-dev libxmu-dev libxi-dev libgl1-mesa-glx libglu1-mesa libglu1-mesa-dev

Open /etc/modprobe.d/blacklist.conf and add:

blacklist amd76x_edac
blacklist vga16fb
blacklist nouveau
blacklist rivafb
blacklist nvidiafb
blacklist rivatv
sudo apt-get remove --purge nvidia*

When you restart your PC, before loging in, try "Ctrl + Alt + F1" to switch to a text-based login. Try:

sudo service lightdm stop
chmod +x cuda*.run
sudo ./cuda*.run
BLAS

Install a BLAS library such as ATLAS, OpenBLAS or MKL. To install BLAS:

sudo apt-get install libatlas-base-dev 
Python

Install Anaconda Python distribution or install the default Python distribution with numpy, scipy, etc.

MATLAB (optional - needed only if you are planning to use the MATLAB interface)

Install MATLAB using a standard distribution.

1.2 Build the custom Caffe version

Set the path correctly in the Makefile.config. You can rename the Makefile.config.example to Makefile.config, as most common parts are filled already. You may need to change it a bit according to your environment.

After this, in Ubuntu 14.04, try:

make

If there are no error messages, you can then compile and install the Python and Matlab wrappers: To install the MATLAB wrapper (at the moment only for MATLAB versions prior to 2017):

make matcaffe

All done! Try our method!

1.3 Run the demo

First download the model that includes the trained weights from this link into the model folder. Also, change the caffepath variable from code/config.m file accordingly.

The MATLAB script for running the demo on all three tasks (i.e. 2D pose estimation, body part labeling and 3D pose estimation) is demoDMHS.m. The MATLAB script for running the demo on the 3D pose estimation task alone (much faster than demoDMHS.m) is demoDMHS_3D.m. Change the displayMode variable to 1 in order to visualize the results.

Please note that our method requires an image cropped around the bounding box of the person. Also, please validate the scales for the 3D pose estimation and body part labeling tasks for other datasets than Human3.6M.

Contact: alin.popa@imar.ro , mihai.zanfir@imar.ro , cristian.sminchisescu@math.lth.se

Acknowledgments

This work was supported in part by the European Research Council Consolidator grant SEED, CNCS-UEFISCDI under JRP-RO-FR-2014-16, the EU Horizon 2020 Grant #688835 (DE-ENIGMA), and SSF.

Releases

No releases published

Packages

No packages published

Languages