-
-
Notifications
You must be signed in to change notification settings - Fork 88
New cuda #157
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
New cuda #157
Conversation
|
Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR ( |
|
@hmaarrfk sorry for dropping your commits, I just wanted to see what the CI says. I will try to make a proper history once this is passing (if it ever does! :) |
|
Argh, fails locally with |
|
i definitely don't care about my commits. I'm more than happy to see a clean history. |
|
Where is CUDA_HOME defined? I think ti should be set to |
|
I find all the CUDA stuff in |
|
The meta file might need a bit of refreshing. Do look through the deleted lines. Many of them are from a recent refresh when I figured i was hitting build issues. You will also need the patch I recently added. |
|
thanks for the heads up! I will re-add all the red lines, for sure. Or feel free to push to this PR if you want. |
|
seems to be working |
|
Hi! This is the friendly automated conda-forge-linting service. I wanted to let you know that I linted all conda-recipes in your PR ( Here's what I've got... For recipe:
|
|
indeed, I discovered the need for |
|
Also I am wondering about libtensorflow (do we need it compiled with cuda as well?) and wondering about if we need to add Maybe I shall have a look at |
|
Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR ( |
|
It seems like:
|
|
Not much should have changed for the CPU builds, so I am not sure what's going on. My only hunch is that Azure's VMs are just different between the runs ... or they decreased the amount of available RAM? |
|
Nice it seems that 2 builds pass. All seems normal on the CPU side! |
|
Are you able to do the uploads to the conda-forge channel? I believe you've addressed all the comments from others. The next step to me is to:
If you can upload, then we can likely move forward. |
|
Argh, sorry, did this get lost again -.- :) Yes, I can do both -- uploads to conda-forge channel (I am a core member) and gather builds for Linux. |
|
hmm that is a good point. I'll let you decide what you want to do. after this PR is done, we can work toward: |
|
@wolfv would you be willing to work @183amir to test his CIs #144 (comment) |
|
Builds are running! :) |
|
He has OSX-amd64 at least :/ Better than nothing. |
|
I am not running osx-arm64 cross compilation, it's running on a native machine and they fail: https://gitlab.idiap.ch/bob/conda/-/jobs/248772 |
|
So, I've finished all the builds (details in the collapsed section below). I am happy to upload these builds to the conda-forge channel (@xhochy / @h-vetinari can you merge & give the go on that?) @183amir happy to help in the other PR. I can also do some tests on a M1 to see what's going on. @hmaarrfk can you help with rebasing #144 once this one here is merged?! All build outputs!Variant: linux_64_cuda_compiler_version10.2cudnn7python3.7.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_version10.2cudnn7python3.8.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_version10.2cudnn7python3.9.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_version11.0cudnn8python3.7.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_version11.0cudnn8python3.8.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_version11.0cudnn8python3.9.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_version11.1cudnn8python3.7.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_version11.1cudnn8python3.8.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_version11.1cudnn8python3.9.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_version11.2cudnn8python3.7.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_version11.2cudnn8python3.8.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_version11.2cudnn8python3.9.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_versionNonecudnnundefinedpython3.7.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_versionNonecudnnundefinedpython3.8.____cpython
SHA256 sums Variant: linux_64_cuda_compiler_versionNonecudnnundefinedpython3.9.____cpython
SHA256 sums |
Thanks a lot for all the builds, and for pu(bli)shing the playbook! I already LGTM'd the changes (aside from one open nit that I don't feel strongly about), but @xhochy still had an open review point (at least he hasn't responded to the comments since his original request), and I'm not sure if someone like @isuruf wants to chime in perhaps. So I'll hold off from merging just yet and wait a bit more if there's gonna be more feedback. |
|
I can rebase #144. I've tried to centralize the open issues in the top comment. |
|
For the record, they seem to use a pretty broad definition of CVEs, including segfaults, crashes & null pointer exceptions. The descriptions of the CVEs don't seem to be public yet, nor their severity ratings. All in all I don't consider this a blocker, if we prepare #144 quickly after merging this. |
|
Ok, I check that tensorflow is actually the last package for both the migrations in #144. We should finish this, solve the CVEs in 144. Otherwise we will be here for a while. It is somewhat challenging to rebase, so I'm worried that we will lose progress if we try to bite off too much. |
xhochy
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We probably want CPU packages preferred over GPU ones as in other packages where we are using track_features.
|
|
||
| cp cc_toolchain_config.bzl cc_toolchain_build_config.bzl | ||
| apply_cc_template cc_toolchain_config.bzl | ||
| apply_cc_template crosstool_wrapper_driver_is_not_gcc |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Fine with merging here but we should follow up on this soon.
|
|
||
| # weigh down cpu implementation and give cuda preference | ||
| track_features: | ||
| - tensorflow-cpu # [cuda_compiler_version == "None"] |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We did do this the other way around normally, esp @isuruf pointed out that we should give CPU builds a higher priority.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I am following pytorch here ... https://github.com/conda-forge/pytorch-cpu-feedstock/blob/b8824904342b8ea142d297668bf0aa32790091a6/recipe/meta.yaml#L148-L149
The pip package also gives the GPU version by default.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Also you'd only get the GPU version if there is actually a working CUDA driver installation on the system (the __cuda virtual package is required by cudatoolkit). In that case I think we can argue that it's also desired to get the GPU version by default.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Makes sense, I would still like to have @isuruf chime in here.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
+1 for weighing down CPU, preferring GPU (where possible)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks!
|
@wolfv Feel free to merge and copy the builds to conda-forge! |
|
I am uploading the GPU builds now, first to my own channel, then to conda-forge! |
|
Builds are copied to the conda-forge channel: Thanks everyone! |
|
I concur, thanks everyone! 🥳 |
Checklist
0(if the version changed)conda-smithy(Use the phrase@conda-forge-admin, please rerenderin a comment in this PR for automated rerendering)Open discussion issues: