-
Notifications
You must be signed in to change notification settings - Fork 72
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
loss or NME computing for convergence #26
Comments
感谢做了好多工作,简单回答以上几个问题: |
对不起,我的评论写的是英文, 中文部分基本是翻译的,可能说的不清楚。 用过pretrained_model, 但是loss 依然在30 以上。而且跳动比较大,不是单线下降的。 |
提的几个问题都是很好的 |
@github-luffy 我直接跑demo,loss也是下降不低于30,不知是不是正常情况,这是一个batch累计的loss? |
@epoc88 效果怎样吗? |
I have tried both with "pretrained_model/mobileNetV2_0.25.pth" and without this pretrained model for retraining the model for common, challenge and fullset.
However, losses never goes down to below 30, neither from pretrained or retrained from initial-scratch.
On the other hand, I tried to removed the training part of the code from train_model.py
see this
https://github.com/epoc88/PFLD_68pts_Pytorch/blob/master/test_NME.py
I did not add any pretrained model, but I still got some NME result..
Test epochs: 3 Loss 0.094
mean error and failure rate
mean error : 0.061
failure rate: L1 0.161
So I am a bit confused...
The issue could be related to optimizer and gradient decent., or loss function WingLoss is good for fine tuning, but PFLF's MSE could be good in the beginning.
Here is a Chinese paper discussing the performance of PFLD on 300W, especially the loss function
https://jishuin.proginn.com/p/763bfbd29621
One more thing, it is worth to mention that the default code for training is "common subset" for validation during training (i.e. without ibug images). There are common, challenge, fullset as validation subset.
我已经尝试了使用“pretrained_model/mobileNetV2_0.25.pth”和没有此预训练模型的重新训练模型,以进行普通、挑战和全集的重新训练。
然而,损失从不低于30 ,无论是从最初的训练或再训练。
另一方面,我试图从train_model.py中删除代码的训练部分
看看这个
https://github.com/epoc88/PFLD_68pts_Pytorch/blob/master/test_NME.py
我没有添加任何预先训练的模型,但我仍然得到了一些NME的结果。
测试阶段:3损耗0.094
平均误差和故障率
平均误差: 0.061
失效率:L1 0.161
所以我有点困惑...
这个问题可能与损耗函数 和 梯度下降法有关。**
下文是一篇关于300W的PFLD性能,特别是损耗函数的中文论文。
https://jishuin.proginn.com/p/763bfbd29621
这是我用您的代码修改后训练的MobileNetV2_0.25模型来测算的结果
主要差别,论文速度1.25ms, 我们的是200ms, MobileV2_0.25 的尺寸也不一样,论文是2.1MB, 我们pretrained_model 是1.1MB, 这个不是官方的pretrained model?
还有一点,值得一提的是,训练的默认代码是训练期间验证的“公共子集”(即。没有ibug图片)。有共同的,挑战的,全集作为验证子集。
The text was updated successfully, but these errors were encountered: