Skip to content

Conversation

petrex
Copy link
Contributor

@petrex petrex commented Mar 21, 2024

The PR

  • fixes int_mm() /int8_gemm() integration with hipblasLT backend (require ROCm 6.0).
  • enables/fixes the following tests on Rocm
    • test__int_mm_k_16_n_16_use_transpose_a_False_use_transpose_b_False_cuda
    • test__int_mm_k_16_n_16_use_transpose_a_False_use_transpose_b_True_cuda
    • test__int_mm_k_16_n_16_use_transpose_a_True_use_transpose_b_False_cuda
    • test__int_mm_k_16_n_16_use_transpose_a_True_use_transpose_b_True_cuda
    • test__int_mm_k_16_n_32_use_transpose_a_False_use_transpose_b_False_cuda
    • test__int_mm_k_16_n_32_use_transpose_a_False_use_transpose_b_True_cuda
    • test__int_mm_k_16_n_32_use_transpose_a_True_use_transpose_b_False_cuda
    • test__int_mm_k_16_n_32_use_transpose_a_True_use_transpose_b_True_cuda
    • test__int_mm_k_32_n_16_use_transpose_a_False_use_transpose_b_False_cuda
    • test__int_mm_k_32_n_16_use_transpose_a_False_use_transpose_b_True_cuda
    • test__int_mm_k_32_n_16_use_transpose_a_True_use_transpose_b_False_cuda
    • test__int_mm_k_32_n_16_use_transpose_a_True_use_transpose_b_True_cuda
    • test__int_mm_k_32_n_32_use_transpose_a_False_use_transpose_b_False_cuda
    • test__int_mm_k_32_n_32_use_transpose_a_False_use_transpose_b_True_cuda
    • test__int_mm_k_32_n_32_use_transpose_a_True_use_transpose_b_False_cuda
    • test__int_mm_k_32_n_32_use_transpose_a_True_use_transpose_b_True_cuda

cc @jeffdaily @sunway513 @jithunnair-amd @pruthvistony @ROCmSupport @dllehr-amd @jataylo @hongxiayang

Copy link

pytorch-bot bot commented Mar 21, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/122431

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit 89ca666 with merge base a046606 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@pytorch-bot pytorch-bot bot added module: rocm AMD GPU support for Pytorch release notes: linalg_frontend release notes category labels Mar 21, 2024
@jeffdaily jeffdaily added the ciflow/rocm Trigger "default" config CI on ROCm label Mar 21, 2024
@ezyang ezyang added the triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module label Mar 22, 2024
@pruthvistony pruthvistony added the rocm This tag is for PRs from ROCm team label Mar 25, 2024
@jithunnair-amd jithunnair-amd requested a review from malfet March 28, 2024 21:33
@soulitzer
Copy link
Contributor

looks okay to me, but will let others do final review

@soulitzer soulitzer removed their request for review April 1, 2024 17:31
@jithunnair-amd
Copy link
Collaborator

@malfet Please review. The changes are ROCm-specific

@pruthvistony
Copy link
Collaborator

@pruthvistony pruthvistony added the rocm priority high priority ROCm PRs from performance or other aspects label Apr 18, 2024
@pruthvistony
Copy link
Collaborator

@pytorchbot rebase

@pytorchmergebot
Copy link
Collaborator

@pytorchbot started a rebase job onto refs/remotes/origin/viable/strict. Check the current status here

@pytorchmergebot
Copy link
Collaborator

Successfully rebased int_mm onto refs/remotes/origin/viable/strict, please pull locally before adding more changes (for example, via git checkout int_mm && git pull --rebase)

@malfet
Copy link
Contributor

malfet commented Apr 23, 2024

@pytorchbot merge

@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Apr 23, 2024
@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

pytorchmergebot pushed a commit that referenced this pull request Apr 24, 2024
Update ROCm-triton to use the AMD backend from https://github.com/openai/triton

Note: `test__int_mm` can be enabled after #122431 is landed

Co-authored-by: Pruthvi Madugundu <pruthvigithub@gmail.com>
Co-authored-by: Nikita Shulga <2453524+malfet@users.noreply.github.com>
Pull Request resolved: #121801
Approved by: https://github.com/nmacchioni, https://github.com/malfet
pytorchmergebot pushed a commit to xuhancn/pytorch that referenced this pull request Apr 24, 2024
Update ROCm-triton to use the AMD backend from https://github.com/openai/triton

Note: `test__int_mm` can be enabled after pytorch#122431 is landed

Co-authored-by: Pruthvi Madugundu <pruthvigithub@gmail.com>
Co-authored-by: Nikita Shulga <2453524+malfet@users.noreply.github.com>
Pull Request resolved: pytorch#121801
Approved by: https://github.com/nmacchioni, https://github.com/malfet
pytorchmergebot pushed a commit that referenced this pull request Apr 25, 2024
Update ROCm-triton to use the AMD backend from https://github.com/openai/triton

Note: `test__int_mm` can be enabled after #122431 is landed

Co-authored-by: Pruthvi Madugundu <pruthvigithub@gmail.com>
Co-authored-by: Nikita Shulga <2453524+malfet@users.noreply.github.com>
Pull Request resolved: #121801
Approved by: https://github.com/nmacchioni, https://github.com/malfet
pruthvistony added a commit to ROCm/pytorch that referenced this pull request Apr 26, 2024
Update ROCm-triton to use the AMD backend from https://github.com/openai/triton

Note: `test__int_mm` can be enabled after pytorch#122431 is landed

Co-authored-by: Pruthvi Madugundu <pruthvigithub@gmail.com>
Co-authored-by: Nikita Shulga <2453524+malfet@users.noreply.github.com>
Pull Request resolved: pytorch#121801
Approved by: https://github.com/nmacchioni, https://github.com/malfet
petrex added a commit to petrex/pytorch that referenced this pull request May 3, 2024
The PR

- fixes int_mm() /int8_gemm() integration with hipblasLT backend (require ROCm 6.0).
- enables/fixes the following tests on Rocm
    - test__int_mm_k_16_n_16_use_transpose_a_False_use_transpose_b_False_cuda
    - test__int_mm_k_16_n_16_use_transpose_a_False_use_transpose_b_True_cuda
    - test__int_mm_k_16_n_16_use_transpose_a_True_use_transpose_b_False_cuda
    - test__int_mm_k_16_n_16_use_transpose_a_True_use_transpose_b_True_cuda
    - test__int_mm_k_16_n_32_use_transpose_a_False_use_transpose_b_False_cuda
    - test__int_mm_k_16_n_32_use_transpose_a_False_use_transpose_b_True_cuda
    - test__int_mm_k_16_n_32_use_transpose_a_True_use_transpose_b_False_cuda
    - test__int_mm_k_16_n_32_use_transpose_a_True_use_transpose_b_True_cuda
    - test__int_mm_k_32_n_16_use_transpose_a_False_use_transpose_b_False_cuda
    - test__int_mm_k_32_n_16_use_transpose_a_False_use_transpose_b_True_cuda
    - test__int_mm_k_32_n_16_use_transpose_a_True_use_transpose_b_False_cuda
    - test__int_mm_k_32_n_16_use_transpose_a_True_use_transpose_b_True_cuda
    - test__int_mm_k_32_n_32_use_transpose_a_False_use_transpose_b_False_cuda
    - test__int_mm_k_32_n_32_use_transpose_a_False_use_transpose_b_True_cuda
    - test__int_mm_k_32_n_32_use_transpose_a_True_use_transpose_b_False_cuda
    - test__int_mm_k_32_n_32_use_transpose_a_True_use_transpose_b_True_cuda

Pull Request resolved: pytorch#122431
Approved by: https://github.com/pruthvistony, https://github.com/jithunnair-amd, https://github.com/malfet, https://github.com/atalman
petrex pushed a commit to petrex/pytorch that referenced this pull request May 3, 2024
Update ROCm-triton to use the AMD backend from https://github.com/openai/triton

Note: `test__int_mm` can be enabled after pytorch#122431 is landed

Co-authored-by: Pruthvi Madugundu <pruthvigithub@gmail.com>
Co-authored-by: Nikita Shulga <2453524+malfet@users.noreply.github.com>
Pull Request resolved: pytorch#121801
Approved by: https://github.com/nmacchioni, https://github.com/malfet
pytorch-bot bot pushed a commit that referenced this pull request May 3, 2024
Update ROCm-triton to use the AMD backend from https://github.com/openai/triton

Note: `test__int_mm` can be enabled after #122431 is landed

Co-authored-by: Pruthvi Madugundu <pruthvigithub@gmail.com>
Co-authored-by: Nikita Shulga <2453524+malfet@users.noreply.github.com>
Pull Request resolved: #121801
Approved by: https://github.com/nmacchioni, https://github.com/malfet
pytorchmergebot pushed a commit that referenced this pull request Oct 7, 2024
This pull request enables the int_mm_error tests for rocm 6.0+ . since  #122431 landed

Pull Request resolved: #124999
Approved by: https://github.com/jeffdaily, https://github.com/malfet
jithunnair-amd added a commit to ROCm/pytorch that referenced this pull request Oct 23, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ciflow/rocm Trigger "default" config CI on ROCm ciflow/trunk Trigger trunk jobs on your pull request Merged module: rocm AMD GPU support for Pytorch open source release notes: linalg_frontend release notes category rocm priority high priority ROCm PRs from performance or other aspects rocm This tag is for PRs from ROCm team triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module

Projects

None yet

Development

Successfully merging this pull request may close these issues.

10 participants