Skip to content

cardboard-q/pifuhd_demo_model_test

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

pifuhd_demo_model_test

Their citation:

@inproceedings{saito2020pifuhd, title={PIFuHD: Multi-Level Pixel-Aligned Implicit Function for High-Resolution 3D Human Digitization}, author={Saito, Shunsuke and Simon, Tomas and Saragih, Jason and Joo, Hanbyul}, booktitle={CVPR}, year={2020} }

Results

Caveats

  • Did this using stock photos (https://www.pexels.com/)
  • Didn't do any pre-processing other than picking photos that I thought could be segmented decently
  • It's running inside a colab notebook, so it has a resolution limit
  • Image quality was lost on the conversion from mp4 to gif.

image image image image image image

Thoughts:

  • I think 512 x 512 is a necessary size for the spatial processsing, so your image will be rescaled to fit
    • since it rescales, square images are best, which I eventually realized
    • you can see how the rescaling can warp the reference images
  • It does the image segmentation for you, so having a clean background helps
    • The README also says that for better results you can pre-process with a segmentation mask, so this makes sense
  • It seems better models resulted from face-on images. Could be the training, or it could be the input. However, my assumption is that you want to see the face and infer the rest (and not the other way around because of facial expressions), so I'm personally fine if it only works on face-on images.
  • The yoga picture catches the leg that is held up, but it also seems to be inferring that there is another leg that is obfuscated. It doesn't help that the colors are pretty uniform in that region.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published