[Inductor] Fallback to super().get_read_writes when epilogue_fusion_user_defined_triton_kernel is disabled#176832
[Inductor] Fallback to super().get_read_writes when epilogue_fusion_user_defined_triton_kernel is disabled#176832coufon wants to merge 1 commit intopytorch:mainfrom
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/176832
Note: Links to docs will display an error until the docs builds have been completed. ✅ You can merge normally! (4 Unrelated Failures)As of commit 904933b with merge base 82374f2 ( FLAKY - The following job failed but was likely due to flakiness present on trunk:
BROKEN TRUNK - The following jobs failed but was present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
UNSTABLE - The following job is marked as unstable, possibly due to flakiness on trunk:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This PR needs a
|
…ser_defined_triton_kernel is disabled (pytorch#176832) Summary: PR pytorch#173662 introduces a new fusion `epilogue_fusion_user_defined_triton_kernel`. It overrides the `get_read_writes` method of `UserDefinedTritonKernel`, even when `epilogue_fusion_user_defined_triton_kernel` is disabled. This may cause regression to existing Triton kernel CUDA graph for other models. This diff fixed the regression by falling back to the original `get_read_writes` when `epilogue_fusion_user_defined_triton_kernel` is disabled Reviewed By: AmesingFlank Differential Revision: D95727036
…ser_defined_triton_kernel is disabled (pytorch#176832) Summary: PR pytorch#173662 introduces a new fusion `epilogue_fusion_user_defined_triton_kernel`. It overrides the `get_read_writes` method of `UserDefinedTritonKernel`, even when `epilogue_fusion_user_defined_triton_kernel` is disabled. This may cause regression to existing Triton kernel CUDA graph for other models. This diff fixed the regression by falling back to the original `get_read_writes` when `epilogue_fusion_user_defined_triton_kernel` is disabled Reviewed By: AmesingFlank Differential Revision: D95727036
…ser_defined_triton_kernel is disabled (pytorch#176832) Summary: Pull Request resolved: pytorch#176832 PR pytorch#173662 introduces a new fusion `epilogue_fusion_user_defined_triton_kernel`. It overrides the `get_read_writes` method of `UserDefinedTritonKernel`, even when `epilogue_fusion_user_defined_triton_kernel` is disabled. This may cause regression to existing Triton kernel CUDA graph for other models. This diff fixed the regression by falling back to the original `get_read_writes` when `epilogue_fusion_user_defined_triton_kernel` is disabled Reviewed By: AmesingFlank Differential Revision: D95727036
|
@pytorchmergebot merge |
Merge failedReason: This PR needs a If not, please add the To add a label, you can comment to pytorchbot, for example For more information, see Details for Dev Infra teamRaised by workflow job |
This PR needs a
|
|
@pytorchmergebot merge |
Merge startedYour change will be merged once all checks pass (ETA 0-4 Hours). Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
The inductor NameError flake (buf1 not defined) that prompted the pin has been fixed upstream by multiple PyTorch PRs: - pytorch/pytorch#176772 (fix is_unfusable scheduler logic) - pytorch/pytorch#176832 (guard get_read_writes behind config flag) - pytorch/pytorch#177062 (fix MultiOutput write deps) The root cause was pytorch/pytorch#173662 which overrode UserDefinedTritonKernel.get_read_writes() even when the epilogue fusion feature was disabled, breaking buffer scheduling for TritonTemplateBuffer subclasses like Helion's HelionTemplateBuffer. Verified the previously-failing test passes against current PyTorch main. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
The inductor NameError flake (buf1 not defined) that prompted the pin has been fixed upstream by multiple PyTorch PRs: - pytorch/pytorch#176772 (fix is_unfusable scheduler logic) - pytorch/pytorch#176832 (guard get_read_writes behind config flag) - pytorch/pytorch#177062 (fix MultiOutput write deps) The root cause was pytorch/pytorch#173662 which overrode UserDefinedTritonKernel.get_read_writes() even when the epilogue fusion feature was disabled, breaking buffer scheduling for TritonTemplateBuffer subclasses like Helion's HelionTemplateBuffer. Verified the previously-failing test passes against current PyTorch main. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
The inductor NameError flake (buf1 not defined) that prompted the pin has been fixed upstream by multiple PyTorch PRs: - pytorch/pytorch#176772 (fix is_unfusable scheduler logic) - pytorch/pytorch#176832 (guard get_read_writes behind config flag) - pytorch/pytorch#177062 (fix MultiOutput write deps) The root cause was pytorch/pytorch#173662 which overrode UserDefinedTritonKernel.get_read_writes() even when the epilogue fusion feature was disabled, breaking buffer scheduling for TritonTemplateBuffer subclasses like Helion's HelionTemplateBuffer. Verified the previously-failing test passes against current PyTorch main. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
…ser_defined_triton_kernel is disabled (pytorch#176832) Summary: PR pytorch#173662 introduces a new fusion `epilogue_fusion_user_defined_triton_kernel`. It overrides the `get_read_writes` method of `UserDefinedTritonKernel`, even when `epilogue_fusion_user_defined_triton_kernel` is disabled. This may cause regression to existing Triton kernel CUDA graph for other models. This diff fixed the regression by falling back to the original `get_read_writes` when `epilogue_fusion_user_defined_triton_kernel` is disabled Differential Revision: D95727036 Pull Request resolved: pytorch#176832 Approved by: https://github.com/AmesingFlank
Summary: PR #173662 introduces a new fusion
epilogue_fusion_user_defined_triton_kernel. It overrides theget_read_writesmethod ofUserDefinedTritonKernel, even whenepilogue_fusion_user_defined_triton_kernelis disabled. This may cause regression to existing Triton kernel CUDA graph for other models. This diff fixed the regression by falling back to the originalget_read_writeswhenepilogue_fusion_user_defined_triton_kernelis disabledDifferential Revision: D95727036
cc @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @kadeng @muchulee8 @amjames @chauhang @aakhundov @coconutruben @jataylo