Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llama7b也有相同的效果吗? #2

Open
JiahengZeng opened this issue Aug 3, 2023 · 0 comments
Open

llama7b也有相同的效果吗? #2

JiahengZeng opened this issue Aug 3, 2023 · 0 comments

Comments

@JiahengZeng
Copy link

我运行了llama7b转中文的强化学习版本,训练过程中可以观察到对一些问题对回答转换比较快,但是对另外一些训练效果不太好。(已经达到25epoch)README中的案例是MOSS的RLHF效果吗?
另:显存在训练中一直增加,torch.cuda.empty_cache()貌似没有很好地起到效果。

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant