-
Notifications
You must be signed in to change notification settings - Fork 22.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
RuntimeError: cuDNN error: CUDNN_STATUS_NOT_SUPPORTED. This error may appear if you passed in a non-contiguous input #32564
Comments
Maybe related to #32395. |
For now, pytorch 1.4 nightly seems to have solved this problem! Thanks! |
Previous suggestions in this thread did not resolve my problem; currently on pytorch-nightly (1.6.0.dev20200525) w/ cuda 10.1.243. Oddly reducing my batch size from 64 to 36 worked 🤷♂️ |
It happens to me as well with It happens when I reach a batch normalization layer with a huge batch size, but when I decrease the batch size the error is gone. It is probably a memory issue that happens when a batch is too big. |
It happens to me as well with |
maybe input size too large |
Just reduce the batch size and try again. It works for me. |
Thanks, reducing the batch_size is OK. But it seems a bug? |
I'm getting this with pytorch 1.6.0. Definitely seems like a bug. Reducing batch size is not a good workaround. Getting the right batch size is critical to certain algorithms & loss functions, such as when doing negative sampling for contrastive learning. Perhaps related, sometimes my code hits this error instead |
It also happens to me |
reduce batch works for me |
Why is this closed? Reducing the batch size does not solve it for me. It is still a bug. |
I was experiencing this problem as well from pytorch=1.4.0... upgrading to pytorch=1.5 solved the issue 😄 |
Thanks for your suggestion. It resolves my issue. |
totally agreed. We ALWAYS need more batches |
It looks like OOM causes this problem too |
add this line after
|
Seems like most of the issues regarding this subject matter is closed. So I am opening this issue again.
Issue description
I am facing this issue when using loss.backward() on my loss function.
Some of the loss function classes I am using are the following.
class CELoss_auxilary(nn.Module):
class MaskedMSELoss(nn.Module):
def init(self):
super(MaskedMSELoss, self).init()
I also checked with other loss functions.
System Info
Pytorch 1.4 cuda-toolkit 10.1 ubuntu 16.0.4.
I am facing this problem only during backward computation in training. My evaluation code runs fine.
Interestingly my code runs fine with this combination:
pytorch1.3 cuda-toolkit 10.0
pytorch1.1 cuda-toolkit 9.0
But I need to use the aforementioned combination pytorch 1.4 cuda-toolkit 10.1 for accessing some sparse convolution tools which are only available in CUDA 10.1>= higher. Can anyone help in this regard?
The text was updated successfully, but these errors were encountered: