-
Notifications
You must be signed in to change notification settings - Fork 25.4k
[CI] Set PATH during build to include location of sccache wrapped nvcc #155464
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/155464
Note: Links to docs will display an error until the docs builds have been completed. ⏳ No Failures, 3 PendingAs of commit 9619911 with merge base 73220d5 ( UNSTABLE - The following job is marked as unstable, possibly due to flakiness on trunk:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
ddc29b6
to
c5427d8
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is some dark magic, please create an issue and let's try to get to the bottom of it....
I.e. unwrapped nvcc should always be in /usr/local/cuda/bin/nvcc
folder, if it's not there, than we are doing something really wrong
The unwrapped version is in |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Good catch! It's true that nvcc
is at /usr/local/cuda/bin/nvcc
and not the sccache-wrapper one. The surprising thing, although, is that cache hit rate looks decent https://github.com/pytorch/pytorch/actions/runs/15562870942/job/43819461677#step:16:472, maybe nvcc stats is not even show there?
@pytorchbot merge -f "im impatient" |
Merge startedYour change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes). Please use Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
Sccache wasn't working for nvcc on jammy, so manually set the path to include where nvcc is
I had problems with always making nvcc a wrapper in some inductor tests where I got
and I also got an error (only on clang) when trying to set CMAKE_CUDA_COMPILER_LAUNCHER to /opt/cache/bin/sccache or sccache
Non jammy cuda jobs' docker images used a different dockerfile, which set CMAKE_CUDA_COMPILER_LAUNCHER
pytorch/.ci/docker/ubuntu-cuda/Dockerfile
Line 110 in e895e96
Alt solution:
Given that I only get the error on clang, I could set CMAKE_CUDA_COMPILER_LAUNCHER=sccache only when not using clang
Setting CUDA_NVCC_EXECUTABLE doesn't fail but also doesn't result in cache hits/misses