We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
需要训练多少epoch是否有个经验值?是否和数据量有关,有没有什么参考意见?
The text was updated successfully, but these errors were encountered:
从经验上看,数据量越大需要训练的epoch越少。 数据如果比较小的话,可以训练数十个epoch。 因为代码里包含了每个epoch保存模型的功能,所以可以对于每个节点分别进行生成,观察并比较生成的效果。 如有1G以上的数据,训练一个epoch其实就可以观察到比较可观的生成效果了。当然,对大数据来说,训练的越久越好。
Sorry, something went wrong.
非常感谢!
试验5个epoch基本可以收敛,英伟达训练megatron gpt2也是5个epoch收敛
No branches or pull requests
需要训练多少epoch是否有个经验值?是否和数据量有关,有没有什么参考意见?
The text was updated successfully, but these errors were encountered: