-
Notifications
You must be signed in to change notification settings - Fork 22.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Version 1.3 no longer supporting Tesla K40m? #30532
Comments
Just to be sure, were you using |
1.3.1
Was the env at point of failure. |
cc @ngimel |
K40m has a compute capability of 3.5, which I believe we have dropped support of. |
Ok. Please may you implement a useful "oldgpu" warning? Like here: #6529 Error at the moment very unclear to casual user like me. --- EDIT ---
Struggl(ed/ing) to find both of those things! As an aside, @ssnl - possibly this line needs updating if you are correct: Line 10 in bf61405
|
@JamesOwers If I'm not mistaken, this commit bumped the minimal compute capability to 3.7. |
There's no technical reason for it to be changed to 3.7 right? This is just for Conda? Looks like it went from 3.5 and 5.0+ to 3.7 and 5.0+ so it was always missing either 3.5 or 3.7. I suppose it takes too long/becomes too large to support more than 2 built architectures. |
@soumith might correct me, but I think the main reason is the growing size of the binaries. |
@ptrblck that is the reason but it is strange it went from supporting K40 (+ several consumer cards) and not K80 to supporting K80 and not K40 (+ several consumer cards).
I also wish there was a way for the message to reflect the minimum cuda arch from the cuda arch list for when it was compiled. This would make it easier when it gets changed to 3.7, for example. Or when a user supports 3.0 by compiling it themselves. |
This is also being discussed at #24205 (comment) |
I'd just like to suggest that the compatible compute capabilities for the precompiled binaries be added somewhere to the documentation, especially when providing installation instructions for the binaries. That information does not appear to be readily available anywhere. |
k40m with cuda10.0 get the same error!!! |
hi guys, i have made a python 3.6 pytorch 1.3.1 linux_x86_64 wheel without restriction on compute capability, and it's working on my 3.5 GPU. i would be more than happy to build wheels for different python and pytorch versions if someone can tell me a proper distribution channel (i.e. not google drive). |
@jayenashar Are you able to provide instruction to build Pytorch version 1.3.1 for a specific GPU (NVIDIA Tesla K20 GPU) & Python 3.6.8? I've attempted to build a compatible version but am still having hardware compatibility issues: |
@anowlan123 I don't see a reason to build for a specific GPU, but I believe you can export the environment variable The pytorch 1.3.1 wheel I made should work for you (python 3.6.9, NVIDIA Tesla K20 GPU). I setup a pypi account to try and distribute it, but it seems there is a 60MB limit, and my wheel is 139MB. So I have uploaded it here: https://github.com/UNSWComputing/pytorch/releases/download/v1.3.1/torch-1.3.1-cp36-cp36m-linux_x86_64.whl |
Dear @anowlan123, I would be very interested in a wheel of pytorch1.4 that work with Keppler K40 and cuda9.2. Would you be able to help out? I am thinking about installing that via miniconda. |
@PeteKey you didn't specify a python version, but i made a wheel with python 3.6, pytorch 1.4.1, and magma-cuda92. please try it here: https://github.com/UNSWComputing/pytorch/releases/download/v1.4.1/torch-1.4.1-cp36-cp36m-linux_x86_64.whl if you have any issues, please upgrade to cuda10.2. |
@anowlan123, python 3.6 is fine but I guess something is not quite working yet. Any idea how to fix this? I am running this on ubuntu 14.04 if that matters. File "/home/pk/miniconda3/envs/pytorch1.4py36_unsw_anowlan123/lib/python3.6/site-packages/torch/init.py", line 81, in |
@jayenashar Thanks, still having the same compatibility issues though. @PeteKey once i create an conda enviroment, i used this script to build from source. #!/bin/bash conda install numpy ninja pyyaml mkl mkl-include setuptools cmake cffi cd ~/anaconda3/envs/env/compiler_compat Prep Pyorch Repocd /home/user/Downloads Specify environment variables for specific pytorch buildexport CMAKE_PREFIX_PATH=${CONDA_PREFIX:-"$(dirname $(which conda))/../"} Build and installpython setup.py install Clean the buildsetup.py clean --all cd ~/anaconda3/envs/env/compiler_compat |
@nelson-liu that's great. then i only need to worry about conda packages. this is how i test for cuda: |
@jayenashar |
@Guptajakala yes i will release to that forked repo, unless someone knows a better place. i tried pypi but it seems they have a file size limit and that is the reason the official builds don't support old GPUs. i can try an anaconda channel. right now i'm taking requests here as it seems to be the discoverable place. |
@jayenashar After installation, I import it but it says conda list shows this item |
@Guptajakala no you can't use a py3.7 pytorch with python 3.6. do you want me to build you one for python 3.6? |
@jayenashar |
@jayenashar aweseome! |
@nelson-liu thank you very much for the wheels, the 1.6 wheel helped me a lot to get a project running just now @jayenashar also thanks to you for being so incredible considerate! if you are still taking requests, could I ask you to please create a build for the following?
|
@igor-krawczuk https://github.com/UNSWComputing/pytorch/releases/tag/v1.7.0 sorry i didn't realise there is a 1.7.1. let me know if you absolutely need it. |
@jayenashar Thank you very much for your availability, I have a Nvidia 920m with cuda capabilities 3.5, can I use your prebuilt packages?
PS: If I use windows could I have some installation problems? Because I tried to install it but after the installation it says |
hi @AlbertoZandara any CUDA GPU will work with my packages. pytorch remove the old compute capabilities to save space and that's why my packages are much larger. unfortunately i don't have windows, so i am only building linux packages: #30532 (comment) |
Hi @jayenashar, any chance I can get a 1.8.0 wheel for the following? PyTorch: 1.8.0 Thanks so much for your help! |
hi @sophiaas unfortunately the way i used to build isn't working anymore. i'll try and get them for you another way, hopefully in the next week, but no promises. |
@jayenashar appreciate you! |
@sophiaas i regret to inform you that i can't figure it out. i can help you to build from source, maybe? |
Hi @jayenashar, any chance I can get a wheel for the most advanced version of PyTorch (>= 1.5) for the following configurations Tesla K40c Thank you so much i'm becoming desperate to make this work |
@geemk i built some but they aren't working. you can try them if they magically work for you: |
@jayenashar Thank you very much for your availability, I have a Nvidia 820m which compute capability is only 2.1, can I get a possible wheel for the following
Thanks so much for your help! |
@beifengli looks like CUDA 9.1 was never supported and 9.0 was last supported in pytorch 1.1. Have you tried https://pytorch.org/get-started/previous-versions/#v110 ? |
@jramseyer I try to use pytorch 1.0 and 1.1, but it always show me the following information: I found that pytorch from version 0.2 recommand GPU compute capability >= 3.0. But my GPU is too old and it's compute capability is only 2.1. my test code:
I really want to use my GPU, I hope you can help, thank you very much. |
sorry i don't think i can help you. 0.1.12 seems like your best chance |
@jayenashar thank you very much. I will try it right away |
Pytorch+GPU **silently** fails with the previous tutorial. It means that with a simple command similar to this "torch.cuda.is_available()" it will show the GPU. It would even "allocate" a tensor to the device with a command similar to this "y = torch.tensor([1,4,9]).to(device)". However, when doing more advanced commands such as ".forward()" or even matrices operations it would print an error similar to this "RuntimeError: CUDA error: no kernel image is available for execution on the device" . The following command above fixes the issue. The pip command installed pytorch in this directory "/users/kfotso/.conda/envs/compat_gpu/lib/python3.7/site-packages/" . **More info here**: --> https://blog.nelsonliu.me/2020/10/13/newer-pytorch-binaries-for-older-gpus/ and here pytorch/pytorch#30532 . **Package can be found here**: --> https://github.com/nelson-liu/pytorch-manylinux-binaries/releases I just downloaded it from here.
Looks like your error message for this is buggy:
It would be helpful if you actually filled in the I strongly suggest updating your main website page (https://pytorch.org/get-started/locally/) to clearly state the minimum CUDA capability index supported. You link to https://developer.nvidia.com/cuda-zone, which gives people the false impression that any CUDA-capable Nvidia GPU will work, which is not the case. |
Agreed, a CUDA capability matrix of Pytorch on the official website would be very helpful. |
🐛 Bug
I am using a Tesla K40m, installed pytorch 1.3 with conda, using CUDA 10.1
To Reproduce
Steps to reproduce the behavior:
conda install pytorch cudatoolkit -c pytorch
.forward()
First tried downgrading to cudatoolkit=10.0, that exhibited same issue.
The code will run fine if you repeat steps above but instead
conda install pytorch=1.2 cudatoolkit=10.0 -c pytorch
.Expected behavior
If no longer supporting a specific GPU, please bomb out upon load with useful error message.
Environment
Unfort ran your script after I 'fixed' so pytorch version will be 1.2 here - issue encountered with version 1.3.
cc @ezyang @gchanan @zou3519 @jerryzh168 @ngimel
The text was updated successfully, but these errors were encountered: