-
Notifications
You must be signed in to change notification settings - Fork 5.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
使用crf层,在多线程GPU下,如果batch_size不为1会出错 #9261
Labels
User
用于标记用户问题
Comments
@jacquesqiao 正在看您的问题 |
processing |
是否有结论,有结论辛苦告知一下。 |
问题已确认,原因是crf层没有gpu kernel,所以做了自动内存转化,这个机制在加了正则之后出现了问题,正在思考解决办法 |
你好,关于上面这个问题。我有一个疑问是如果Place选择是GPU的单个卡,这时候ParallelDo是否还会起作用,是否存在一个GPU下的多线程,还说我上面在一个GPU卡下进行多线程本身是错的?在多个卡上起多线程应该怎么做? |
已解决。 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
发现的问题是使用包含crf层的代码时,多线程(ParallelDo)在单个gpu下,如果batch_size不为1(比如10),会出文末粘贴的问题。值的注意的是,这和#9234 的错误是一致的。
使用的代码是:
这是对Paddle/python/paddle/fluid/tests/book/test_label_semantic_roles.py的一个多线程改写的尝试。通过将test_label_semantic_roles.py变成多线程来复现这个问题,也可以联系我提供环境。
The text was updated successfully, but these errors were encountered: