Skip to content

ROCm enable sparse_sampled_addmm#86401

Closed
jpvillam-amd wants to merge 3 commits intopytorch:masterfrom
ROCm:ROCm-SDDMM
Closed

ROCm enable sparse_sampled_addmm#86401
jpvillam-amd wants to merge 3 commits intopytorch:masterfrom
ROCm:ROCm-SDDMM

Conversation

@jpvillam-amd
Copy link
Contributor

@jpvillam-amd jpvillam-amd commented Oct 6, 2022

Enables:
test_comprehensive_sparse_sampled_addmm_cuda_complex128
test_comprehensive_sparse_sampled_addmm_cuda_complex64
test_comprehensive_sparse_sampled_addmm_cuda_float32
test_comprehensive_sparse_sampled_addmm_cuda_float64
test_dispatch_meta_sparse_sampled_addmm_cuda_complex128
test_dispatch_meta_sparse_sampled_addmm_cuda_complex64
test_dispatch_meta_sparse_sampled_addmm_cuda_float32
test_dispatch_meta_sparse_sampled_addmm_cuda_float64
test_meta_sparse_sampled_addmm_cuda_complex128
test_meta_sparse_sampled_addmm_cuda_complex64
test_meta_sparse_sampled_addmm_cuda_float32
test_meta_sparse_sampled_addmm_cuda_float64

cc @jeffdaily @sunway513 @jithunnair-amd @ROCmSupport

@pytorch-bot
Copy link

pytorch-bot bot commented Oct 6, 2022

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/86401

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures, 1 Pending

As of commit ddac27b:
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

Copy link
Collaborator

@ngimel ngimel left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Lint errors are real, otherwise lgtm

@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Oct 6, 2022
@jpvillam-amd jpvillam-amd changed the title Added hip version checking and hipification code ROCm enable sparse_sampled_addmm Oct 6, 2022
@pytorch-bot pytorch-bot bot added the module: rocm AMD GPU support for Pytorch label Oct 6, 2022
@jpvillam-amd
Copy link
Contributor Author

    test_comprehensive_sparse_sampled_addmm_cuda_complex128 (__main__.TestDecompCUDA) ... ok (0.187s)
    test_comprehensive_sparse_sampled_addmm_cuda_complex64 (__main__.TestDecompCUDA) ... ok (0.091s)
    test_comprehensive_sparse_sampled_addmm_cuda_float32 (__main__.TestDecompCUDA) ... ok (0.072s)
    test_comprehensive_sparse_sampled_addmm_cuda_float64 (__main__.TestDecompCUDA) ... ok (0.074s)

From: https://github.com/pytorch/pytorch/actions/runs/3201111231/jobs/5228949868#step:8:22000

    test_dispatch_meta_sparse_sampled_addmm_cuda_complex128 (__main__.TestMetaCUDA) ... ok (0.029s)
    test_dispatch_meta_sparse_sampled_addmm_cuda_complex64 (__main__.TestMetaCUDA) ... ok (0.010s)
    test_dispatch_meta_sparse_sampled_addmm_cuda_float32 (__main__.TestMetaCUDA) ... ok (0.009s)
    test_dispatch_meta_sparse_sampled_addmm_cuda_float64 (__main__.TestMetaCUDA) ... ok (0.009s)

From: https://github.com/pytorch/pytorch/actions/runs/3201111231/jobs/5228949868#step:8:6521

    test_meta_sparse_sampled_addmm_cuda_complex128 (__main__.TestMetaCUDA) ... ok (0.010s)
    test_meta_sparse_sampled_addmm_cuda_complex64 (__main__.TestMetaCUDA) ... ok (0.010s)
    test_meta_sparse_sampled_addmm_cuda_float32 (__main__.TestMetaCUDA) ... ok (0.009s)
    test_meta_sparse_sampled_addmm_cuda_float64 (__main__.TestMetaCUDA) ... ok (0.010s)

From: https://github.com/pytorch/pytorch/actions/runs/3201111231/jobs/5228949868#step:8:11747

@jpvillam-amd
Copy link
Contributor Author

I think the 1 failed check is from a XML reporting tool segfault


OK (skipped=12, expected failures=2)

Generating XML reports...
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutograd-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-autograd.test_complex.TestAutogradComplex-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutogradDeviceTypeCUDA-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutogradForwardMode-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutogradForwardModeBatchedGrad-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-autograd.test_functional.TestAutogradFunctional-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutogradInferenceMode-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestAutogradMultipleDispatchCUDA-20221007003900.xml
Generated XML report: test-reports/python-unittest/test_autograd/TEST-TestMultithreadAutograd-20221007003900.xml
[db1a429837ff:11768:0:11842] Caught signal 11 (Segmentation fault: address not mapped to object at address 0x9)

FINISHED PRINTING LOG FILE of test_autograd (/var/lib/jenkins/workspace/test/test-reports/test_autograd_u5hojsgw)

Traceback (most recent call last):
  File "/var/lib/jenkins/workspace/test/run_test.py", line 1236, in <module>
    main()
  File "/var/lib/jenkins/workspace/test/run_test.py", line 1211, in main
    raise RuntimeError(err_message)
RuntimeError: test_autograd failed! Received signal: SIGSEGV

Don't think that is related to this PR. @jithunnair-amd

@jithunnair-amd
Copy link
Collaborator

@pytorchbot rebase

@pytorchmergebot
Copy link
Collaborator

@pytorchbot successfully started a rebase job. Check the current status here

@pytorchmergebot
Copy link
Collaborator

Successfully rebased ROCm-SDDMM onto refs/remotes/origin/viable/strict, please pull locally before adding more changes (for example, via git checkout ROCm-SDDMM && git pull --rebase)

@jithunnair-amd
Copy link
Collaborator

@pytorchbot merge

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

kulinseth pushed a commit to kulinseth/pytorch that referenced this pull request Nov 5, 2022
Enables:
test_comprehensive_sparse_sampled_addmm_cuda_complex128
test_comprehensive_sparse_sampled_addmm_cuda_complex64
test_comprehensive_sparse_sampled_addmm_cuda_float32
test_comprehensive_sparse_sampled_addmm_cuda_float64
test_dispatch_meta_sparse_sampled_addmm_cuda_complex128
test_dispatch_meta_sparse_sampled_addmm_cuda_complex64
test_dispatch_meta_sparse_sampled_addmm_cuda_float32
test_dispatch_meta_sparse_sampled_addmm_cuda_float64
test_meta_sparse_sampled_addmm_cuda_complex128
test_meta_sparse_sampled_addmm_cuda_complex64
test_meta_sparse_sampled_addmm_cuda_float32
test_meta_sparse_sampled_addmm_cuda_float64

Pull Request resolved: pytorch#86401
Approved by: https://github.com/ngimel
kulinseth pushed a commit to kulinseth/pytorch that referenced this pull request Dec 10, 2022
Enables:
test_comprehensive_sparse_sampled_addmm_cuda_complex128
test_comprehensive_sparse_sampled_addmm_cuda_complex64
test_comprehensive_sparse_sampled_addmm_cuda_float32
test_comprehensive_sparse_sampled_addmm_cuda_float64
test_dispatch_meta_sparse_sampled_addmm_cuda_complex128
test_dispatch_meta_sparse_sampled_addmm_cuda_complex64
test_dispatch_meta_sparse_sampled_addmm_cuda_float32
test_dispatch_meta_sparse_sampled_addmm_cuda_float64
test_meta_sparse_sampled_addmm_cuda_complex128
test_meta_sparse_sampled_addmm_cuda_complex64
test_meta_sparse_sampled_addmm_cuda_float32
test_meta_sparse_sampled_addmm_cuda_float64

Pull Request resolved: pytorch#86401
Approved by: https://github.com/ngimel
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ciflow/trunk Trigger trunk jobs on your pull request cla signed Merged module: rocm AMD GPU support for Pytorch open source release notes: sparse release notes category

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants