Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

求教一般训练几个epoch? #16

Closed
ZhaoyueSun opened this issue Aug 13, 2019 · 3 comments
Closed

求教一般训练几个epoch? #16

ZhaoyueSun opened this issue Aug 13, 2019 · 3 comments

Comments

@ZhaoyueSun
Copy link

需要训练多少epoch是否有个经验值?是否和数据量有关,有没有什么参考意见?

@Morizeyao
Copy link
Owner

从经验上看,数据量越大需要训练的epoch越少。
数据如果比较小的话,可以训练数十个epoch。
因为代码里包含了每个epoch保存模型的功能,所以可以对于每个节点分别进行生成,观察并比较生成的效果。
如有1G以上的数据,训练一个epoch其实就可以观察到比较可观的生成效果了。当然,对大数据来说,训练的越久越好。

@ZhaoyueSun
Copy link
Author

非常感谢!

@fengzuo97
Copy link

试验5个epoch基本可以收敛,英伟达训练megatron gpt2也是5个epoch收敛

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants