-
-
Notifications
You must be signed in to change notification settings - Fork 180
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Are you going to release the trained SAN model on 300w? #15
Comments
Hi Roberto, thanks for trying our codes. Would you mind to let me know how did you get these numbers? What normalization distance did you use? The released model is trained by using |
Hi, I have processed 300W using the trained model that you provide called SAN_300W_GTB_itn_cpm_3_50_sigma4_128x128x8 I use inter-pupil distance normalization as in the literature. As a result, I attach some examples of images that I have obtained with your model. Are these images correct? (NME is showed at the bottom left of each image). Are these images similar to the ones that you have obtained in your paper? It is impossible to obtain a 3.98 NME in the Full data set with the model that I am using (https://github.com/D-X-Y/SAN#evaluation-on-the-single-image). I look forward to your response. Best regards, |
Hi, I just updated the readme, you can follow https://github.com/D-X-Y/SAN#evaluate-on-300-w to evaluate the released model on 300-W. I think there may be two reasons that cause the worse results from your evaluation:
Note about the normalization distance: We follow "A deep regression architecture with two-stage re-initialization for high performance facial landmark detection, CVPR 2017" and "300 faces in-the-wild challenge: The first facial landmark localization challenge, ICCV-W 2013" |
I am using the ground truth face bounding box from the 300W annotations. I have changed the normalization measure to corners distance. The reported results are not the same that you report in the paper. However, they are closer to the 3.98 mentioned before.
However, I encourage you to modify Table 1 in your paper because literature (SDM, ESR, LBF, CFSS) does not use corners normalization either. Finally, could you provide some example of images of 300W with your prediction and the NME obtained? On the other hand, are you going to release the trained SAN model on AFLW? |
If you use Thanks for your suggestion about modifing Table 1. Before we submit our paper, we didn't notice that SDM, ESR, LBF, CFSS are using inter-pupil, but simply copying the numbers from "A deep regression architecture with two-stage re-initialization for high performance facial landmark detection". That is our mistake. After several months of the CVRP camera ready, we noticed this mistake but can not change that version. I have updated this information in our README and will clarify it in our following papers. You can refer Figure 8 in the paper for examples. I'm reaching some deadlines, and can not provide NME for specific examples right now. For the trained SAN model on AFLW, if you want to reproduce the results, you can run commands following |
Dear Xuanyi Dong,
First of all I would like to congratulate you for your excellent work. I'm a PhD student at Spain. My research is focused on face alignment. I have used your https://github.com/D-X-Y/SAN code and I would like to ask some questions:
We would like to repeat the 3.98 NME obtained in the Full set on 300W. I look forward to your response.
Best regards,
Roberto Valle
The text was updated successfully, but these errors were encountered: