-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Clarification on Pose to Body #8
Comments
|
Can you pease elaborate on the steps of the workflow for the pose-to-body application? |
Both DensePose and OpenPose will generate a (3-channel) color image out of the box. I simply concatenated these two images together, forming a 6-channel input to the network. |
@dustinfreeman haven't been able to get either to work for Pose2Body, getting CUDA issues as described in #32 |
@tcwang0509 Hi, it's an amazing work, regarding to pose to body task, during my training, I found the face of dancer was the most difficult part to generate using your model, is there any trick you applied to optimise face exclusively? like the one used in the paper "everybody dance now"? |
Regarding face, do you observe quality degrade on training or test images? If it's training, there's an additional face discriminator (by |
@tcwang0509 many thanks for your suggestions, very useful, now it's much better |
Excuse me? When I use the option -add_face_disc, I got an error: Traceback (most recent call last): Do you know why and how to fix it? Thanks in advance~ |
Good news! I've just found the solution. It is because that when you want to add the face discriminator, the number of patch scales of the face discriminator will be set to "num_D - 2" according to the code: |
@pshah123 in regards to your questions about multiple pose estimations, did you work this out and how to control it? I have a model containing multiple people and it would be great to be able to interpolate between them, wondered if you had any ideas? Maybe some sort to z-vector to control the generator when testing, at the moment it just goes to whichever training person was closest to the densepose/openpose data. With the paper you released a video of multiple output for face to edge, are these contained within one model (or 3 different models) and if so how do you control which one it generates? |
@dustinfreeman @tcwang0509 @cuuupid @jakeelwes @tsing90 vid2vid# This post is about the discussion of pose2body, and my statement here is appropriate. |
I have a few questions regarding the pose --> body task. From the paper,
By "directly concatenate" do you also layer the DensePose UV pose on the OpenPose color-coded pose? Or just use one of the two, and if so which performs better?
The provided example shows a still background for the dancer. If I understand correctly would a video with changing background or multiple people cause issues? e.g. the demonstration for DensePose and OpenPose includes a fast paced multi-person dance video. However the poses generated are all more or less synchronized, and the background is not encoded in any way. Would the model generate mostly noise for the background in this case, and would it be able to synthesize human bodies from multiple pose estimations?
How does the test set perform against other dancers? Do changes such as height, limb length, etc. cause issues in generation?
Lastly, is there a dockerized version of this repository available? Alternatively, would this model compile inside the Flownet2 Docker container?
Thanks!
The text was updated successfully, but these errors were encountered: