Skip to content

hiroyasuakada/3D-Human-Pose-Perception-from-Egocentric-Stereo-Videos

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

36 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

3D-Human-Pose-Perception-from-Egocentric-Stereo-Videos (CVPR 2024)

The official PyTorch inference code of our CVPR 2024 paper, "3D Human Pose Perception from Egocentric Stereo Videos".

img

For any questions, please contact the first author, Hiroyasu Akada [hakada@mpi-inf.mpg.de] .

[Project Page] [Dataset Page]

Citation

@inproceedings{hakada2024unrealego2,
  title = {3D Human Pose Perception from Egocentric Stereo Videos},
  author = {Akada, Hiroyasu and Wang, Jian and Golyanik, Vladislav and Theobalt, Christian},
  booktitle = {Computer Vision and Pattern Recognition (CVPR)},
  year = {2024}
}

Updates

  • 12/12/2025: We decided to release the test split of UnrealEgo2 and UnrealEgo-RW to facilitate the field of egocentric 3D vision.

UnrealEgo2/UnrealEgo-RW Datasets

Download

You can download the UnrealEgo2/UnrealEgo-RW datasets on our dataset page.

[12 Dec, 2025]: We decided to release the test split of UnrealEgo2 and UnrealEgo-RW to facilitate the field of egocentric 3D vision!

Depths from SfM/Metashape

You can download depth data from SfM/Metashape, as described in our paper.

Note that these depth data differ from the synthetic pixel-perfect depth maps available on our benchmark challenge page.

Implementation

Dependencies

We tested our code with the following dependencies:

  • Python 3.9
  • Ubuntu 18.04
  • PyTorch 2.0.0
  • Cuda 11.7

Please install other dependencies:

pip install -r requirements.txt    

Inference

Trained models

You can download our trained models. Please save them in ./log/(experiment_name).

Inference on UnrealEgo2 test dataset

    bash scripts/test/unrealego2_pose-qa-avg-df_data-ue2_seq5_skip3_B32_lr2-4_pred-seq_local-device_pad.sh

        --data_dir [path to the `UnrealEgoData2_test_rgb` dir]
        --metadata_dir [path to the `UnrealEgoData2_test_sfm` dir]

Please modify the arguments above. The pose predictions will be saved in ./results/UnrealEgoData2_test_pose (raw and zip versions).

Inference on UnrealEgo-RW test dataset

  • Model without pre-training on UnrealEgo2

      bash scripts/test/unrealego2_pose-qa-avg-df_data-ue-rw_seq5_skip3_B32_lr2-4_pred-seq_local-device_pad.sh
    
          --data_dir [path to the `UnrealEgoData_rw_test_rgb` dir]
          --metadata_dir [path to the `UnrealEgoData_rw_test_sfm` dir]
    
  • Model with pre-training on UnrealEgo2

      bash scripts/test/unrealego2_pose-qa-avg-df_data-ue2_seq5_skip3_B32_lr2-4_pred-seq_local-device_pad_finetuning_epoch5-5.sh
    
          --data_dir [path to the `UnrealEgoData_rw_test_rgb` dir]
          --metadata_dir [path to the `UnrealEgoData_rw_test_sfm` dir]
    

Please modify the arguments above. The pose predictions will be saved in ./results/UnrealEgoData_rw_test_pose (raw and zip versions).

Note that UnrealEgo2 is fully compatible with UnrealEgo. This means that you can train your method on UnrealEgo2 and test it on UnrealEgo, and vice versa.

The UnrealEgo dataset (train/validation/test splits) is also publicly available here, including 72 body joint annotations (32 for body and 40 for hand).

About

The official code of our CVPR 2024 paper, "3D Human Pose Perception from Egocentric Stereo Videos".

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors