-
Notifications
You must be signed in to change notification settings - Fork 108
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bart长文本训练问题 #14
Comments
出错的位置好像是在predict位置。当取消在训练过程中进行eval处理时,训练得以正常进行。大佬 Traceback (most recent call last): |
我看看evaluate的逻辑 |
多谢。 我看到另一处地方,应该有些问题: def preprocess_data_bart(data): |
对的,fixed: 7a0be59 |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.(由于长期不活动,机器人自动关闭此问题,如果需要欢迎提问) |
我用Bart训练代码,每个训练数据都为:输入文本约1000字符,输出文本长3-5万字符。训练几个epoch后会出错,错误信息如下所示。
但是控制输入和输出的字符长度,比如都为100字符左右,则训练正常,没有报错。
请问一下:Bart模型的输入输出长度有什么要求吗,这应该是内部embedding维度出错了吧。谢谢。
RuntimeError: CUDA error: CUBLAS_STATUS_EXECUTION_FAILED when calling
cublasSgemmStridedBatched( handle, opa, opb, m, n, k, &alpha, a, lda, stridea, b, ldb, strideb, &beta, c, ldc, stridec, num_batches)
The text was updated successfully, but these errors were encountered: