Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reclaim in cuDNN conv algorithm search #1711

Merged
merged 1 commit into from Jan 2, 2023

Conversation

ToucheSir
Copy link
Contributor

Per discussion on Slack. This dramatically improves the maximum batch size Flux models with conv layers can support without OOMing on GPU, e.g. https://discourse.julialang.org/t/memory-challenges-for-flux-on-resnet/85385.

@CarloLucibello
Copy link
Contributor

Is there something we can do on the NNlibCUDA side while this lands in the next release (I don't know how long that will take but the last tag was in july)?

@ToucheSir
Copy link
Contributor Author

I'm not sure if CUDA.jl does backports, but if so I'd be happy to replicate this PR against any backport branch as well.

@maleadt
Copy link
Member

maleadt commented Jan 2, 2023

I'm not sure if CUDA.jl does backports, but if so I'd be happy to replicate this PR against any backport branch as well.

I've done backports in the past, so I'd be fine doing another backports release.

@maleadt maleadt merged commit 9362065 into JuliaGPU:master Jan 2, 2023
maleadt pushed a commit that referenced this pull request Jan 2, 2023
This dramatically improves the maximum batch size Flux models with conv layers can support without OOMing on GPU, e.g. https://discourse.julialang.org/t/memory-challenges-for-flux-on-resnet/85385.
@ToucheSir ToucheSir deleted the cudnn-algo-reclaim branch January 2, 2023 18:37
maleadt pushed a commit that referenced this pull request Jan 6, 2023
This dramatically improves the maximum batch size Flux models with conv layers can support without OOMing on GPU, e.g. https://discourse.julialang.org/t/memory-challenges-for-flux-on-resnet/85385.
simonbyrne pushed a commit to simonbyrne/CUDA.jl that referenced this pull request Nov 13, 2023
This dramatically improves the maximum batch size Flux models with conv layers can support without OOMing on GPU, e.g. https://discourse.julialang.org/t/memory-challenges-for-flux-on-resnet/85385.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants