Skip to content

[Inductor][CUDA][test] Fix test_mm_plus_mm3_dynamic_shapes_gpu_wrapper on CUDA#175569

Closed
NayanNagabhushana-28 wants to merge 2 commits intopytorch:mainfrom
NayanNagabhushana-28:fix_test_mm_plus_mm3_test
Closed

[Inductor][CUDA][test] Fix test_mm_plus_mm3_dynamic_shapes_gpu_wrapper on CUDA#175569
NayanNagabhushana-28 wants to merge 2 commits intopytorch:mainfrom
NayanNagabhushana-28:fix_test_mm_plus_mm3_test

Conversation

@NayanNagabhushana-28
Copy link
Copy Markdown
Contributor

@NayanNagabhushana-28 NayanNagabhushana-28 commented Feb 23, 2026

test_mm_plus_mm3_gpu_wrapper was marked xfail on CUDA but actually passes, causing XPASS. The expected failure was only in the test_mm_plus_mm3_dynamic_shapes_gpu_wrapper.

Removed xfail from test_mm_plus_mm3 in test/inductor/test_select_algorithm.py and added test_mm_plus_mm3_dynamic_shapes to test_failures_gpu_wrapper list in test/inductor/test_gpu_cpp_wrapper.py

Fixes: #175546

CC: @eqy

cc @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @kadeng @muchulee8 @amjames @chauhang @aakhundov @coconutruben @jataylo

@pytorch-bot
Copy link
Copy Markdown

pytorch-bot Bot commented Feb 23, 2026

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/175569

Note: Links to docs will display an error until the docs builds have been completed.

✅ You can merge normally! (2 Unrelated Failures)

As of commit 2043606 with merge base b9b2033 (image):

FLAKY - The following job failed but was likely due to flakiness present on trunk:

BROKEN TRUNK - The following job failed but was present on the merge base:

👉 Rebase onto the `viable/strict` branch to avoid these failures

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@NayanNagabhushana-28 NayanNagabhushana-28 changed the title Fix test_mm_plus_mm3_dynamic_shapes_gpu_wrapper on CUDA [Inductor][CUDA][test] Fix test_mm_plus_mm3_dynamic_shapes_gpu_wrapper on CUDA Feb 23, 2026
@soulitzer soulitzer requested review from eellison and jansel February 27, 2026 22:25
@soulitzer soulitzer added the triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module label Feb 27, 2026
@jansel jansel added ciflow/inductor ciflow/trunk Trigger trunk jobs on your pull request labels Feb 28, 2026
@jansel
Copy link
Copy Markdown
Contributor

jansel commented Feb 28, 2026

@pytorchbot merge -i

@pytorchmergebot
Copy link
Copy Markdown
Collaborator

Merge started

Your change will be merged while ignoring the following 0 checks:

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

pytorchmergebot pushed a commit that referenced this pull request Mar 6, 2026
# Motivation
This PR aims to fix the CI failure on XPU.
- `test_mm_plus_mm3` seems to only fail on CUDA. Already fixed in #175569
- `test_codegen_with_custom_heuristics_module` will fail on XPU due to `ModuleNotFoundError: No module named 'extension_triton_heuristics'`. This is because on CUDA CI, `is_parallel` is `False`, and on XPU CI, it is `True` due to a race condition. So we should add the path to `sys.path` in the parent process so that the` ExtensionCachingAutotuner` class can be resolved to boost the UT's robustness, whatever `is_parallel` it is.
- skip `test_circular_dependencies` due to it is flaky on XPU, see #110040

https://github.com/pytorch/pytorch/blob/a88bb129e9d9e7572bc3a830ad5d148d74a63c48/torch/_inductor/async_compile.py#L385

# Additional Context
fix #173473
fix #173344
fix #173916
fix #110040

Pull Request resolved: #176057
Approved by: https://github.com/jansel
EmanueleCoradin pushed a commit to EmanueleCoradin/pytorch that referenced this pull request Mar 30, 2026
…r on CUDA (pytorch#175569)

`test_mm_plus_mm3_gpu_wrapper` was marked xfail on CUDA but actually passes, causing XPASS. The expected failure was only in the `test_mm_plus_mm3_dynamic_shapes_gpu_wrapper`.

Removed xfail from `test_mm_plus_mm3` in `test/inductor/test_select_algorithm.py` and added `test_mm_plus_mm3_dynamic_shapes` to `test_failures_gpu_wrapper` list in `test/inductor/test_gpu_cpp_wrapper.py`

Fixes: pytorch#175546

CC: @eqy

Pull Request resolved: pytorch#175569
Approved by: https://github.com/jansel
EmanueleCoradin pushed a commit to EmanueleCoradin/pytorch that referenced this pull request Mar 30, 2026
# Motivation
This PR aims to fix the CI failure on XPU.
- `test_mm_plus_mm3` seems to only fail on CUDA. Already fixed in pytorch#175569
- `test_codegen_with_custom_heuristics_module` will fail on XPU due to `ModuleNotFoundError: No module named 'extension_triton_heuristics'`. This is because on CUDA CI, `is_parallel` is `False`, and on XPU CI, it is `True` due to a race condition. So we should add the path to `sys.path` in the parent process so that the` ExtensionCachingAutotuner` class can be resolved to boost the UT's robustness, whatever `is_parallel` it is.
- skip `test_circular_dependencies` due to it is flaky on XPU, see pytorch#110040

https://github.com/pytorch/pytorch/blob/a88bb129e9d9e7572bc3a830ad5d148d74a63c48/torch/_inductor/async_compile.py#L385

# Additional Context
fix pytorch#173473
fix pytorch#173344
fix pytorch#173916
fix pytorch#110040

Pull Request resolved: pytorch#176057
Approved by: https://github.com/jansel
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ciflow/inductor ciflow/trunk Trigger trunk jobs on your pull request Merged module: inductor open source topic: not user facing topic category triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[Inductor][CUDA][test] test_mm_plus_mm3_gpu_wrapper fails with XPASS on CUDA.

5 participants