Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

请教一下OOM问题 #52

Closed
beeper00 opened this issue Oct 15, 2021 · 2 comments
Closed

请教一下OOM问题 #52

beeper00 opened this issue Oct 15, 2021 · 2 comments

Comments

@beeper00
Copy link

用自己造的数据跑模型时,由于部分句子较长,容易出现oom问题,所以我在代码中加了句子长度不超过15的限制(11g显存)才能正常训练。默认的batchsize是2048,我改这个数字发现实际使用的显存没有变化,似乎不起作用。想请教一下,如果不想限制句子长度,应该改哪部分参数或代码来解决oom?

@zhangyimi
Copy link
Collaborator

请问使用的是什么模型? 建议模型选用ernie-lstm或者transformer,batch_size先设置为300试试。

@beeper00
Copy link
Author

beeper00 commented Oct 15, 2021 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants