New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Demo #9
Comments
When I am free, I will add a demo for inference. |
@MassyMeniche @leoxiaobin With a quick dive into codes, find a function to get max preds:
Which What does that function gets? How to get the final keypoints coordinates finally? |
@jinfagang Have you solved this problem yet? |
@jinfagang , after you get the preds, you should also need to project the coordinates to the original image, using the function at https://github.com/leoxiaobin/deep-high-resolution-net.pytorch/blob/master/lib/core/inference.py#L49**** |
@leoxiaobin What does the center and scale mean? |
what does the center and scale mean?... |
@njustczr After a digging in, I think it's the object detection box.. which means you should do object detection first.. |
center:bbox center? scale: the ratio of (width / height) ? |
get_max_preds() performs better than get_final_preds()?... scale=height/200.0 |
I have the same question. Please share how did you get keypoints on your own data. |
By refecence this code[https://github.com/microsoft/human-pose-estimation.pytorch/issues/26#issuecomment-447404791], I can get good result.
The command is: |
@Ixiaohuihuihui Hi, I test imgs use your codes. However the render results is bad. Do you know the reason of this cases? Thanks. |
I don't know the special issuses. But I think maybe we should modify the parameters according to datasets. Which dataset images did you test on? |
@lxiaohuihuihui thanks so much for sharing! If I try with Coco data I get perfect results, but not on my own (because they are scaled differently?). Could you maybe explain what scale and pixel_std refer to exactly? I guess this is where it goes wrong. Thanks in advance! |
Please refer: microsoft/human-pose-estimation.pytorch#26 (comment) Actually, I also don't know how to test in the wild image elegantly, but I guess maybe you can get the parameter by drawing a detection box manually or using fasterrcnn to detect the people. Your image size should be consistent wit the reference image in coco. |
I actually use detection bounding boxes from Mask RCNN and with coco data it works, I also checked wether the bboxes are correct and they are. Thanks anyway :) |
As mentioned in microsoft/human-pose-estimation.pytorch#26 (comment) the error was due to this line: c, s = _box2cs(box, data_numpy.shape[0], data_numpy.shape[1]) instead it should be: c, s = _box2cs(box, data_numpy.shape[1], data_numpy.shape[0]) So image_width and image_height were basically switched. I guess it worked better for Coco data, because the images are a lot more symmetrical than mine. |
@leoxiaobin @lxiaohuihuihui @MassyMeniche @jinfagang @wait1988 @njustczr
|
@tengshaofeng I'm not getting perfect results on your data either. I used w48_384x288 on your data but for example if I try on random data, where you can see the body parts better it works: |
@carlottaruppert , thanks so much for your reply. I think your performance is better than mine. Have u used the operation of fliping when test? |
Have you done this? It is essantial. I haven't used flipping in testing, I am using a slightly altered version of the script posted in this issue. |
@carlottaruppert ,yes , I try as you said. It performance better. Really thanks for your advice. |
I think my result is better... since the picture width and height is basically the size of the bbox I scipped the Mask RCNN and coded the bbox hard. In addition I made sure that this part is commented out: I think it's confused by the dress again, so the legs aren't good |
@carlottaruppert , when i comment the "if center[0] != -1: |
@carlottaruppert , when i set the box as large as the image, and I use the w48_384x288,but the result is as follows, I do not know why I can not get your result. |
from future import absolute_import
def _box2cs(box, image_width, image_height): def _xywh2cs(x, y, w, h, image_width, image_height):
def main():
if name == 'main': Maybe your box format is not as it should be (x, y, width, height)? This is the version without the Mask RCNN annotation reading. and I'm calling it like this: |
if you are using Mask RCNN as well, change the bbox format with this function: def change_box_to_coco_format(mask_box):
|
@carlottaruppert , thanks so much. |
@carlottaruppert , I found the problem. because of the bbox, my bbox is [0, 0, 130, 410], yours is |
@carlottaruppert , I have found the solution after I read the code carefully. Actually,it shoud be: |
@tengshaofeng thank you so much! You're right! Don't know how I could miss that and no idea why it almost worked for me... |
@carlottaruppert hello, can i ask you some details? |
@eng100200 sure, just ask. |
@carlottaruppert how many datasets you used in training? I want to train for multi-person in indoor environment. |
@carlottaruppert thanks for your reply |
@eng100200 I didn't train at all. I'm only using HRNet to label my data. Sorry, but I guess I cannot help you. |
@carlottaruppert so you have used the test code only. Which pre-trained model you have used coco or mpii? |
I used the test code, to check wether retina or mask rcnn works better as a human detector for HR net and now I'm simply using the demo code. I am using coco pre trained weights w48. |
@carlottaruppert ok, can you share your email with me? My email is sm_adnan21@hotmail.com |
I would say this issue can be closed with #161 being merged |
@Ixiaohuihuihui But it shows
The command is:
Do you know the reason for this case? Thanks. |
First of all thank you for the great work.Is it possible to directly input the intercepted body image into the demo function?The predited picture I get is poor,I do not konw why. |
why pixel_std is equal to 200? |
I just did what is stated above but I am getting the issue as : |
Thank you, I have received the e-mail! I will reply as soon as possible——Linhui Dai
|
First of all thank you for the great work. I'm currently trying to set up a demo of the estimator but run into some issues in the post-processing stage (the network output is a B x 17 x 128 x 128 for 512x512 images)
Are planning to release any helper functions for post-processing the output to a key-points ?
Many thanks
The text was updated successfully, but these errors were encountered: