Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

[Feature Request] Cache the CUDNN convolution optimization result #10567

Open
ThomasDelteil opened this issue Apr 16, 2018 · 9 comments
Open

Comments

@ThomasDelteil
Copy link
Contributor

ThomasDelteil commented Apr 16, 2018

On every script run, I get the CUDNN convolution optimization algorithm running. This can take a few seconds, I wonder if we could cache the result locally based on a hash of MXNet + CUDA + CUDNN version for each device ID (or whatever could cause a change in algorithm selection) ?

[20:48:19] src/operator/nn/./cudnn/./cudnn_algoreg-inl.h:107: Running performance tests to find the best convolution algorithm, this can take a while... (setting env variable MXNET_CUDNN_AUTOTUNE_DEFAULT to 0 to disable
@rajanksin
Copy link
Contributor

@eric-haibin-lin : Please label : CUDA, Feature

@KellenSunderland
Copy link
Contributor

Just want to +1. I've talked to quite a few MXNet users who could really use this functionality.

@1frey
Copy link

1frey commented Sep 19, 2018

+1
Any news?

@KellenSunderland
Copy link
Contributor

My team has an implementation of this in a fork. We'll try and contribute it back, but no promises on a timeline.

@AustinDoolittle
Copy link

Any updates @KellenSunderland? That feature sounds very useful.

@zeryx
Copy link

zeryx commented Mar 8, 2019

@KellenSunderland would love to know more, this is very relevant to us.

@Neutron3529
Copy link
Contributor

+1
2 years has past..

@leezu
Copy link
Contributor

leezu commented Apr 30, 2020

May be fixed as part of cudnn 8 integration? https://docs.nvidia.com/deeplearning/sdk/cudnn-api/index.html

cc @DickJC123

@feevos
Copy link
Contributor

feevos commented Apr 27, 2021

This would be super useful, especially in distributed computing: the time required for optimization, scales badly with number of GPUs/computation nodes.

Thank you for your efforts.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Projects
None yet
Development

No branches or pull requests

10 participants