Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

accuracy on provided model vs facenet paper reported accuracy #35

Closed
nbenhaim opened this issue Jul 23, 2016 · 1 comment
Closed

accuracy on provided model vs facenet paper reported accuracy #35

nbenhaim opened this issue Jul 23, 2016 · 1 comment

Comments

@nbenhaim
Copy link

You report accuracy of 91% on the LFW data. The facenet paper reports 99+. Do you have any insight as to what they did differently vs your trained model? Just curious, thanks

@davidsandberg
Copy link
Owner

Yes, I think this is the key question and I don't have the final answer.
The typical answer is to use a bigger model and feed it more data. However this does not seem to be the solution in this case. The model is not really overfitting, i.e. validation accuracy is not significantly lower than the training accuracy. Also, the performance seems to increase when a smaller (and more shallow) model is used. This points in the direction that the network is not really converging very well. A possible remedy for this is to try for example residual networks (http://arxiv.org/abs/1512.03385). But I don't have any results for this yet.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants