Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ROCm enable sparse_sampled_addmm #86401

Closed
wants to merge 3 commits into from
Closed

Conversation

jpvillam-amd
Copy link
Contributor

@jpvillam-amd jpvillam-amd commented Oct 6, 2022

Enables:
test_comprehensive_sparse_sampled_addmm_cuda_complex128
test_comprehensive_sparse_sampled_addmm_cuda_complex64
test_comprehensive_sparse_sampled_addmm_cuda_float32
test_comprehensive_sparse_sampled_addmm_cuda_float64
test_dispatch_meta_sparse_sampled_addmm_cuda_complex128
test_dispatch_meta_sparse_sampled_addmm_cuda_complex64
test_dispatch_meta_sparse_sampled_addmm_cuda_float32
test_dispatch_meta_sparse_sampled_addmm_cuda_float64
test_meta_sparse_sampled_addmm_cuda_complex128
test_meta_sparse_sampled_addmm_cuda_complex64
test_meta_sparse_sampled_addmm_cuda_float32
test_meta_sparse_sampled_addmm_cuda_float64

cc @jeffdaily @sunway513 @jithunnair-amd @ROCmSupport

@pytorch-bot
Copy link

pytorch-bot bot commented Oct 6, 2022

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/86401

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures, 1 Pending

As of commit ddac27b:
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

Copy link
Collaborator

@ngimel ngimel left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Lint errors are real, otherwise lgtm

@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Oct 6, 2022
@jpvillam-amd jpvillam-amd changed the title Added hip version checking and hipification code ROCm enable sparse_sampled_addmm Oct 6, 2022
@pytorch-bot pytorch-bot bot added the module: rocm AMD GPU support for Pytorch label Oct 6, 2022
@jpvillam-amd
Copy link
Contributor Author

    test_comprehensive_sparse_sampled_addmm_cuda_complex128 (__main__.TestDecompCUDA) ... ok (0.187s)
    test_comprehensive_sparse_sampled_addmm_cuda_complex64 (__main__.TestDecompCUDA) ... ok (0.091s)
    test_comprehensive_sparse_sampled_addmm_cuda_float32 (__main__.TestDecompCUDA) ... ok (0.072s)
    test_comprehensive_sparse_sampled_addmm_cuda_float64 (__main__.TestDecompCUDA) ... ok (0.074s)

From: https://github.com/pytorch/pytorch/actions/runs/3201111231/jobs/5228949868#step:8:22000

    test_dispatch_meta_sparse_sampled_addmm_cuda_complex128 (__main__.TestMetaCUDA) ... ok (0.029s)
    test_dispatch_meta_sparse_sampled_addmm_cuda_complex64 (__main__.TestMetaCUDA) ... ok (0.010s)
    test_dispatch_meta_sparse_sampled_addmm_cuda_float32 (__main__.TestMetaCUDA) ... ok (0.009s)
    test_dispatch_meta_sparse_sampled_addmm_cuda_float64 (__main__.TestMetaCUDA) ... ok (0.009s)

From: https://github.com/pytorch/pytorch/actions/runs/3201111231/jobs/5228949868#step:8:6521

    test_meta_sparse_sampled_addmm_cuda_complex128 (__main__.TestMetaCUDA) ... ok (0.010s)
    test_meta_sparse_sampled_addmm_cuda_complex64 (__main__.TestMetaCUDA) ... ok (0.010s)
    test_meta_sparse_sampled_addmm_cuda_float32 (__main__.TestMetaCUDA) ... ok (0.009s)
    test_meta_sparse_sampled_addmm_cuda_float64 (__main__.TestMetaCUDA) ... ok (0.010s)

From: https://github.com/pytorch/pytorch/actions/runs/3201111231/jobs/5228949868#step:8:11747

@jpvillam-amd
Copy link
Contributor Author

I think the 1 failed check is from a XML reporting tool segfault


OK (skipped=12, expected failures=2)

Generating XML reports...
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutograd-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-autograd.test_complex.TestAutogradComplex-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutogradDeviceTypeCUDA-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutogradForwardMode-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutogradForwardModeBatchedGrad-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-autograd.test_functional.TestAutogradFunctional-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutogradInferenceMode-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutogradMultipleDispatchCUDA-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestMultithreadAutograd-20221007003900.xml
[db1a429837ff:11768:0:11842] Caught signal 11 (Segmentation fault: address not mapped to object at address 0x9)

FINISHED PRINTING LOG FILE of test_autograd (/var/lib/jenkins/workspace/test/test-reports/test_autograd_u5hojsgw)

Traceback (most recent call last):
  File "/var/lib/jenkins/workspace/test/run_test.py", line 1236, in <module>
    main()
  File "/var/lib/jenkins/workspace/test/run_test.py", line 1211, in main
    raise RuntimeError(err_message)
RuntimeError: test_autograd failed! Received signal: SIGSEGV

Don't think that is related to this PR. @jithunnair-amd

@jithunnair-amd
Copy link
Collaborator

@pytorchbot rebase

@pytorchmergebot
Copy link
Collaborator

@pytorchbot successfully started a rebase job. Check the current status here

@pytorchmergebot
Copy link
Collaborator

Successfully rebased ROCm-SDDMM onto refs/remotes/origin/viable/strict, please pull locally before adding more changes (for example, via git checkout ROCm-SDDMM && git pull --rebase)

@jithunnair-amd
Copy link
Collaborator

@pytorchbot merge

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

sgrigory pushed a commit to sgrigory/pytorch that referenced this pull request Oct 28, 2022
Enables:
test_comprehensive_sparse_sampled_addmm_cuda_complex128
test_comprehensive_sparse_sampled_addmm_cuda_complex64
test_comprehensive_sparse_sampled_addmm_cuda_float32
test_comprehensive_sparse_sampled_addmm_cuda_float64
test_dispatch_meta_sparse_sampled_addmm_cuda_complex128
test_dispatch_meta_sparse_sampled_addmm_cuda_complex64
test_dispatch_meta_sparse_sampled_addmm_cuda_float32
test_dispatch_meta_sparse_sampled_addmm_cuda_float64
test_meta_sparse_sampled_addmm_cuda_complex128
test_meta_sparse_sampled_addmm_cuda_complex64
test_meta_sparse_sampled_addmm_cuda_float32
test_meta_sparse_sampled_addmm_cuda_float64

Pull Request resolved: pytorch#86401
Approved by: https://github.com/ngimel
kulinseth pushed a commit to kulinseth/pytorch that referenced this pull request Nov 5, 2022
Enables:
test_comprehensive_sparse_sampled_addmm_cuda_complex128
test_comprehensive_sparse_sampled_addmm_cuda_complex64
test_comprehensive_sparse_sampled_addmm_cuda_float32
test_comprehensive_sparse_sampled_addmm_cuda_float64
test_dispatch_meta_sparse_sampled_addmm_cuda_complex128
test_dispatch_meta_sparse_sampled_addmm_cuda_complex64
test_dispatch_meta_sparse_sampled_addmm_cuda_float32
test_dispatch_meta_sparse_sampled_addmm_cuda_float64
test_meta_sparse_sampled_addmm_cuda_complex128
test_meta_sparse_sampled_addmm_cuda_complex64
test_meta_sparse_sampled_addmm_cuda_float32
test_meta_sparse_sampled_addmm_cuda_float64

Pull Request resolved: pytorch#86401
Approved by: https://github.com/ngimel
kulinseth pushed a commit to kulinseth/pytorch that referenced this pull request Dec 10, 2022
Enables:
test_comprehensive_sparse_sampled_addmm_cuda_complex128
test_comprehensive_sparse_sampled_addmm_cuda_complex64
test_comprehensive_sparse_sampled_addmm_cuda_float32
test_comprehensive_sparse_sampled_addmm_cuda_float64
test_dispatch_meta_sparse_sampled_addmm_cuda_complex128
test_dispatch_meta_sparse_sampled_addmm_cuda_complex64
test_dispatch_meta_sparse_sampled_addmm_cuda_float32
test_dispatch_meta_sparse_sampled_addmm_cuda_float64
test_meta_sparse_sampled_addmm_cuda_complex128
test_meta_sparse_sampled_addmm_cuda_complex64
test_meta_sparse_sampled_addmm_cuda_float32
test_meta_sparse_sampled_addmm_cuda_float64

Pull Request resolved: pytorch#86401
Approved by: https://github.com/ngimel
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ciflow/trunk Trigger trunk jobs on your pull request cla signed Merged module: rocm AMD GPU support for Pytorch open source release notes: sparse release notes category
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

6 participants