New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[ROCm] Custom gfx1100 kernel sample fails to build (clang-offload-bundler not found) #16899
Comments
Yes, I even used the cursed amdgpu-pro installer. ➜ ls /opt/rocm/bin
amdclang amdclang-cpp hipcc hipcc_cmake_linker_helper hipconfig.pl hipdemangleatp hipfc hipvars.pm roc-obj-extract rocm_agent_enumerator
amdclang++ amdflang hipcc.bin hipconfig hipconvertinplace-perl.sh hipexamine-perl.sh hipify-clang offload-arch roc-obj-ls rocminfo
amdclang-cl amdlld hipcc.pl hipconfig.bin hipconvertinplace.sh hipexamine.sh hipify-perl roc-obj rocm-smi |
I think @sogartar faced something similar? |
@raikonenfnu can you first confirm where the clang-offload-bundler binary should be? Do you have it under /opt/rocm like Ben or installed system-wide? |
We may need to check for this during the cmake configuration step. |
I only have it on I also do not have clang-offload-bundler on my env and was able to compile. |
Oh wait you are talking about macrokernel not microkernel, so my previous assumption/comments might be correct here. The previous comments were more about microkernel. I need to check a bit more about samples macrokernel. I think it may be the Perhaps missing a |
@kuhar Was able to repro your issue on my system as well. But if I specify (EDIT: Deleted log from using -nogpublib ) (EDIT: this one actually works if we point to where the PATH=$PATH:/opt/rocm/llvm/bin /home/stanley/nod/iree-build-notrace/llvm-project/bin/clang-19 -x hip --offload-device-only --offload-arch=gfx1100 --rocm-path=/opt/rocm -fuse-cuid=none -O3 /home/stanley/nod/iree/samples/custom_dispatch/hip/kernels/kernels.cu -o /home/stanley/nod/iree-build-notrace/samples/custom_dispatch/hip/kernels/kernels_gfx1100.co |
Thanks, with this set
I set both Separately from solving this, why do we even build this test data in the |
OK it does work after switching from the rocm installation from the amdgpu-pro installer to https://github.com/nod-ai/TheRock/releases/tag/nightly-staging-20240328.41 , setting |
The last remaining issue is the following error:
|
@raikonenfnu @antiagainst should we disable these rocm kernels and make them experimental? They don't seem to work out of the box on a typical linux installation but are included in the main ninja targets |
Ping. This still doesn't build for me. After manually patching the cuda kernel, I'm hitting an issue with another tool missing from path:
Seems like this needs a very specific system-wide installation. |
Error:
My rocm installation is under
/opt/rocm
, the version is 5.7.1.The text was updated successfully, but these errors were encountered: