-
Notifications
You must be signed in to change notification settings - Fork 997
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add support for AMD / ROCm / HIP #707
Comments
AMD is working on it: https://github.com/ROCmSoftwarePlatform/flash-attention I've not tested it yet, but a new branch with WMMA optimizations for Radeon 7000 was added just yesterday it seems. |
I have composed this guide for my AMD AI configuration... Alas it does not work with the ROCm 6 at time of writing. An issue with this is that the AMD fork is always behind What would be helpful is for AMD's content to be included |
@tridao is it possible to merge this to support ROCm? |
I think that's a fork maintained by AMD folks and it's not meant to be merged. |
I doubt they would disapprove of merging, Seems just a rift of communication. I will reach out. |
As it's been a while, and they haven't updated or integrated... It is more likely they offered an example of what could be done - and how to do it, so that the 'community' could take it from there. [ If that's not the case, then they'd clearly mention that, or keep it private. ] I have contacted exllamav2 about the version issue, here is what they said - AMD's offered version isn't of much use... |
Maybe @howiejayz could be part of this conversation =) |
Unfortunately I am no longer working on this project :( But as far as I know the other team is still working on this project and it will be long-term support. |
I want to again request AMD support, since it is now much more popular and usable than it has been
The text was updated successfully, but these errors were encountered: