-
Notifications
You must be signed in to change notification settings - Fork 21.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The support for 3080 or 3090 #45021
Comments
You are using pytorch 1.1, which is not compiled (optimized) for cuda compute capability 8.x (Ampere). Please get the latest CUDA 11. For better performance, please also get the latest pytorch source code, and build them with
Please try the latest pytorch wheel built with cuda 11.x here https://pytorch.org/get-started/locally/ |
Thanks for your reply. I have used your method. However, the problem is |
Seems like the latest cuda 11 doesn't support sm_86 yet. Can you try |
Thanks, I have tried TORCH_CUDA_ARCH_LIST='8.0+PTX'. The problem "nvcc fatal : Unsupported gpu architecture 'compute_86'" is gone. And I compile it successfully. By the way, " a cuda 11 pytorch nightly" also works well. However, a warning indicates: |
I'm wondering when and whether the pytorch/Nvidia official can give support to sm_86? |
Thanks. Are you able to train network on that install? If not, then we may have to wait for the next cuda release that supports sm_86. |
I will try it and give your feedback later. |
Currently, I try: It works well. |
@xwang233 |
@WangWenhao0716 can you try using PyTorch nightly builds (by running |
@malfet OK, in fact, I have downloaded the nightly version Pytorch and used it. The experiments show a significant improvement on multi-GPUs training because I do not compile Pytorch using NCCL by myself. I will report the detailed results soon. |
@malfet Currently, I'm waiting for the arrival of the last GPU. When it arrives, I will report the performance difference of 4 GPUs between 3080 and 1080Ti. |
Will 3090 support memory pooling in pytorch? |
I got the 3090 today. Cuda 11.1 + latest official Torch version does not work. I got "no kernel image is available for execution on the device" error. |
Maybe the nightly build will work but I nuked the system entirely. Anyway, I will get the next 3090 in a few days and test if parallel training (multiple GPUs) work on this nightly version. |
@WangWenhao0716 @xwang233 Do you know if it is safe now to compile with TORCH_CUDA_ARCH_LIST=8.6 with CUDA 11.1? Thanks! |
@bryanhpchiang Hi, now I'm thinking about this question! Therefore, |
It should be fine to build pytorch with I'm not familiar if pytorch release will use cuda 11.1 cc @malfet |
@xwang233 I tried to compile it by myself using cuda11.1. However, this time, I'm facing "fatal error: magma.h: No such file or directory" |
@malfet Hi, I think currently Pytorch does not support cuda11.1. And the performance of 3080 and 3090 is limited. Is that right? When will Pytorch fully support cuda11.1? |
cc @ptrblck for 11.1 perf vs 11.0 perf for 3080 |
3080 gpu can now use pytorch? |
Yes but it does not perform ideally.
发自我的 iPad
…------------------ Original ------------------
From: luobin97 <notifications@github.com>
Date: Sat,Oct 10,2020 4:17 PM
To: pytorch/pytorch <pytorch@noreply.github.com>
Cc: WenhaoWang <2609667791@qq.com>, Mention <mention@noreply.github.com>
Subject: Re: [pytorch/pytorch] The support for 3080 or 3090 (#45021)
3080 gpu can now use pytorch?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub, or unsubscribe.
|
I'm working on cuda 11.1 and cudnn 8.04 |
自己编,没办法 |
but I still got : my cuda version is 11.1 |
I do not think it is possible because it works on my 3090. Please uninstall PyTorch thoroughly and reinstall it. You must point out cudatoolkit=11.0. |
you're right ,I just somehow installed torch1.7 with pip . problem solved ! thank you! 兄弟你太牛逼了! |
hhhhh |
sorry but, it failed again with: RuntimeError: nvrtc: error: invalid value for --gpu-architecture (-arch) |
please add my wechat
Perhaps we can discuss this problem further
My wechat 18602459215
发自我的 iPad
…------------------ Original ------------------
From: wellhowtosay <notifications@github.com>
Date: Wed,Nov 4,2020 0:21 AM
To: pytorch/pytorch <pytorch@noreply.github.com>
Cc: Wenhao Wang <2609667791@qq.com>, Mention <mention@noreply.github.com>
Subject: Re: [pytorch/pytorch] The support for 3080 or 3090 (#45021)
hhhhh
sorry but, it failed again with: RuntimeError: nvrtc: error: invalid value for --gpu-architecture (-arch)
you sure the command works fine with cuda11.1? or maybe the right command is
conda install pytorch torchvision cudatoolkit=11 -c **pytorch-nightly**
which you mentioned above?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub, or unsubscribe.
|
When I ran some code for pytorch it says to goto https://pytorch.org/get-started/locally/ - but there's no mention of 3090 on the screen - so I ended up here.
➜ stylegan2-pytorch git:(master) python convert_weight.py --repo ../stylegan2 stylegan2-ffhq-config-f.pkl UPDATE / just grab the latest nightly build https://pytorch.org/get-started/locally/ pip install numpy
pip install --pre torch torchvision torchaudio -f https://download.pytorch.org/whl/nightly/cu110/torch_nightly.html |
I can't take credit for this, or remember where I found it, but what worked for me was
|
It works! |
Hi! I have got several 3080 GPUs. However, I find some problems. I find the step (1) and (2) very smooth, which means it nearly does not cost any time. |
It is a very old problem. And PyTorch support 3080 and 3090 officially since 1.7.0. |
Greetings, The problem I will note below only happens when I debug with code but not on a python terminal. I believe my setup is properly loaded. Below all info necessary. Error message:
Environment Information: my setup tries to follow the Nvidia compatibility matrix: driver-470/toolkit 114/libcudnn8.2/pytorch1.9+cu111 PyTorch version: 1.9.0+cu111 OS: Ubuntu 20.04.2 LTS (x86_64) Python version: 3.8.10 (default, Jun 2 2021, 10:49:15) [GCC 9.4.0] (64-bit runtime) Nvidia driver version: 470.57.02 Versions of relevant libraries: If I run the code below in a python or python interpreter I get no errors:
It only happen inside Microsoft Code. I have double checked the correct interpreter/branch is loaded... But continue to get the same error. I also tried many other combinations of drivers/toolkits/cudnn/pytorch but have not been able to get this right. It's killing me. Any pointers please? |
" when I debug with code but not on a python terminal." ( I highly recommend miniconda - it integrates within vscode. |
I partly agree with johndpope, You need to check out the environment first
As part of my suggestion. |
Thank you @johndpope and @Heermosi I have continued to study my environment and continue to think it is properly setup in code... Currently, inside
and/or
But if I try to do:
Then, I get the error
Why might passing the device inside the torch call cause such an error but not outside? |
Thank you for the feedback. There were indeed problems in my environment. Packages were not being saved locally but globally... I started the virtual environment from scratch again and verified each installation and software version to ensure consistency. I came up with the following set of instructions in case it might help others:
|
ys.platform: linux
TorchVision: 0.9.1+cu111 the nvidia-smi is +-----------------------------------------------------------------------------+ +-----------------------------------------------------------------------------+ when i trained, i got the RuntimeError: CUDA error: no kernel image is available for execution on the device. |
Please install it using conda or pip, do Not compile by yourself.
发自我的iPhone
…------------------ Original ------------------
From: Max ***@***.***>
Date: Sat,Jan 15,2022 8:57 PM
To: pytorch/pytorch ***@***.***>
Cc: Wenhao Wang ***@***.***>, State change ***@***.***>
Subject: Re: [pytorch/pytorch] The support for 3080 or 3090 (#45021)
sys.platform: linux
Python: 3.6.9 (default, Oct 8 2020, 12:12:24) [GCC 8.4.0]
CUDA available: True
GPU 0: NVIDIA GeForce RTX 3060 Laptop GPU
CUDA_HOME: /usr/local/cuda-11.1
NVCC: Build cuda_11.1.TC455_06.29069683_0
GCC: gcc (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0
PyTorch: 1.8.1+cu111
PyTorch compiling details: PyTorch built with:
GCC 7.3
C++ Version: 201402
Intel(R) Math Kernel Library Version 2020.0.0 Product Build 20191122 for Intel(R) 64 architecture applications
Intel(R) MKL-DNN v1.7.0 (Git Hash 7aed236906b1f7a05c0917e5257a1af05e9ff683)
OpenMP 201511 (a.k.a. OpenMP 4.5)
NNPACK is enabled
CPU capability usage: AVX2
CUDA Runtime 11.1
NVCC architecture flags: -gencode;arch=compute_37,code=sm_37;-gencode;arch=compute_50,code=sm_50;-gencode;arch=compute_60,code=sm_60;-gencode;arch=compute_70,code=sm_70;-gencode;arch=compute_75,code=sm_75;-gencode;arch=compute_80,code=sm_80;-gencode;arch=compute_86,code=sm_86
CuDNN 8.0.5
Magma 2.5.2
Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, CUDA_VERSION=11.1, CUDNN_VERSION=8.0.5, CXX_COMPILER=/opt/rh/devtoolset-7/root/usr/bin/c++, CXX_FLAGS= -Wno-deprecated -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -fopenmp -DNDEBUG -DUSE_KINETO -DUSE_FBGEMM -DUSE_QNNPACK -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -O2 -fPIC -Wno-narrowing -Wall -Wextra -Werror=return-type -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wno-sign-compare -Wno-unused-parameter -Wno-unused-variable -Wno-unused-function -Wno-unused-result -Wno-unused-local-typedefs -Wno-strict-overflow -Wno-strict-aliasing -Wno-error=deprecated-declarations -Wno-stringop-overflow -Wno-psabi -Wno-error=pedantic -Wno-error=redundant-decls -Wno-error=old-style-cast -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, PERF_WITH_AVX512=1, TORCH_VERSION=1.8.1, USE_CUDA=ON, USE_CUDNN=ON, USE_EXCEPTION_PTR=1, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON,
TorchVision: 0.9.1+cu111
OpenCV: 4.5.2
MMCV: 1.2.6
MMCV Compiler: GCC 7.5
MMCV CUDA Compiler: 11.1
MMDetection: 2.9.0
MMDetection3D: 0.10.0+
when i trained, i just got the meesage RuntimeError: CUDA error: no kernel image is available for execution on the device,i tried when i compiled with the TORCH_CUDA_ARCH_LIST=8.6, it is still not working, someone who met this issues??
—
Reply to this email directly, view it on GitHub, or unsubscribe.
Triage notifications on the go with GitHub Mobile for iOS or Android.
You are receiving this because you modified the open/close state.Message ID: ***@***.***>
|
Thanks for replying, do you mean i install that the mmdetection3d using conda or pip? is that right? |
for mmdetection3d please according to its own instruction.
发自我的iPhone
…------------------ Original ------------------
From: Max ***@***.***>
Date: Sat,Jan 15,2022 10:10 PM
To: pytorch/pytorch ***@***.***>
Cc: Wenhao Wang ***@***.***>, State change ***@***.***>
Subject: Re: [pytorch/pytorch] The support for 3080 or 3090 (#45021)
Please install it using conda or pip, do Not compile by yourself. 发自我的iPhone
…
------------------ Original ------------------ From: Max @.> Date: Sat,Jan 15,2022 8:57 PM To: pytorch/pytorch @.> Cc: Wenhao Wang @.>, State change @.> Subject: Re: [pytorch/pytorch] The support for 3080 or 3090 (#45021) sys.platform: linux Python: 3.6.9 (default, Oct 8 2020, 12:12:24) [GCC 8.4.0] CUDA available: True GPU 0: NVIDIA GeForce RTX 3060 Laptop GPU CUDA_HOME: /usr/local/cuda-11.1 NVCC: Build cuda_11.1.TC455_06.29069683_0 GCC: gcc (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0 PyTorch: 1.8.1+cu111 PyTorch compiling details: PyTorch built with: GCC 7.3 C++ Version: 201402 Intel(R) Math Kernel Library Version 2020.0.0 Product Build 20191122 for Intel(R) 64 architecture applications Intel(R) MKL-DNN v1.7.0 (Git Hash 7aed236906b1f7a05c0917e5257a1af05e9ff683) OpenMP 201511 (a.k.a. OpenMP 4.5) NNPACK is enabled CPU capability usage: AVX2 CUDA Runtime 11.1 NVCC architecture flags: -gencode;arch=compute_37,code=sm_37;-gencode;arch=compute_50,code=sm_50;-gencode;arch=compute_60,code=sm_60;-gencode;arch=compute_70,code=sm_70;-gencode;arch=compute_75,code=sm_75;-gencode;arch=compute_80,code=sm_80;-gencode;arch=compute_86,code=sm_86 CuDNN 8.0.5 Magma 2.5.2 Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, CUDA_VERSION=11.1, CUDNN_VERSION=8.0.5, CXX_COMPILER=/opt/rh/devtoolset-7/root/usr/bin/c++, CXX_FLAGS= -Wno-deprecated -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -fopenmp -DNDEBUG -DUSE_KINETO -DUSE_FBGEMM -DUSE_QNNPACK -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -O2 -fPIC -Wno-narrowing -Wall -Wextra -Werror=return-type -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wno-sign-compare -Wno-unused-parameter -Wno-unused-variable -Wno-unused-function -Wno-unused-result -Wno-unused-local-typedefs -Wno-strict-overflow -Wno-strict-aliasing -Wno-error=deprecated-declarations -Wno-stringop-overflow -Wno-psabi -Wno-error=pedantic -Wno-error=redundant-decls -Wno-error=old-style-cast -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, PERF_WITH_AVX512=1, TORCH_VERSION=1.8.1, USE_CUDA=ON, USE_CUDNN=ON, USE_EXCEPTION_PTR=1, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, TorchVision: 0.9.1+cu111 OpenCV: 4.5.2 MMCV: 1.2.6 MMCV Compiler: GCC 7.5 MMCV CUDA Compiler: 11.1 MMDetection: 2.9.0 MMDetection3D: 0.10.0+ when i trained, i just got the meesage RuntimeError: CUDA error: no kernel image is available for execution on the device,i tried when i compiled with the TORCH_CUDA_ARCH_LIST=8.6, it is still not working, someone who met this issues?? — Reply to this email directly, view it on GitHub, or unsubscribe. Triage notifications on the go with GitHub Mobile for iOS or Android. You are receiving this because you modified the open/close state.Message ID: @.***>
Thanks for replying, do you mean i install that the mmdetection3d using conda or pip? is that right?
—
Reply to this email directly, view it on GitHub, or unsubscribe.
Triage notifications on the go with GitHub Mobile for iOS or Android.
You are receiving this because you modified the open/close state.Message ID: ***@***.***>
|
Hi! I have got several 3080 GPUs. However, I find some problems.
The system is Ubuntu 16.04, the version of PyTorch is 1.1.0.
My codes are:
import torch ---(1)
X = torch.rand((3,6)) ---(2)
X = X.cuda() ---(3)
I find the step (1) and (2) very smooth, which means it nearly does not cost any time.
However, I found the third step costs nearly 10 minutes to finish moving the data from memory to VRAM.
I do not know what brings the problem and I also tried other versions of GPU and other 3080s.
I'm looking forward to your reply. Thanks!
cc @malfet @seemethere @walterddr @ngimel
The text was updated successfully, but these errors were encountered: