Skip to content

[Graph Partition] support standalone_compile#154698

Closed
BoyuanFeng wants to merge 3 commits intomainfrom
bf/partition-standalone-compile
Closed

[Graph Partition] support standalone_compile#154698
BoyuanFeng wants to merge 3 commits intomainfrom
bf/partition-standalone-compile

Conversation

@BoyuanFeng
Copy link
Contributor

@BoyuanFeng BoyuanFeng commented May 30, 2025

For graph partition, write_get_raw_stream_header_once is done once so the autotune code may not have the header. This PR additionally calls write_get_raw_stream_header in codegen_device_guard_enter before get_raw_stream is used.

cc @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @chenyang78 @kadeng @muchulee8 @amjames @chauhang @aakhundov

@BoyuanFeng BoyuanFeng added ciflow/trunk Trigger trunk jobs on your pull request topic: not user facing topic category module: inductor labels May 30, 2025
@pytorch-bot
Copy link

pytorch-bot bot commented May 30, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/154698

Note: Links to docs will display an error until the docs builds have been completed.

✅ You can merge normally! (2 Unrelated Failures)

As of commit 15ff86e with merge base 28cb3c0 (image):

BROKEN TRUNK - The following job failed but were present on the merge base:

👉 Rebase onto the `viable/strict` branch to avoid these failures

UNSTABLE - The following job is marked as unstable, possibly due to flakiness on trunk:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

Comment on lines +1599 to +1612
def test_basic(self, device: str, format: str, dynamic: bool) -> None:
self._test_basic(device, format, dynamic)

@config.patch({"fx_graph_cache": True})
@config.patch({"fx_graph_remote_cache": False})
@functorch_config.patch({"enable_autograd_cache": True})
@parametrize("device", (GPU_TYPE, "cpu"))
@parametrize("format", ("binary", "unpacked"))
@parametrize("dynamic", (False, True))
@config.patch("graph_partition", True)
def test_basic_with_graph_partition(
self, device: str, format: str, dynamic: bool
) -> None:
self._test_basic(device, format, dynamic)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

instead of breaking this test into two tests, make graph partition be another boolean parameter, and when use with config.patch()

@BoyuanFeng
Copy link
Contributor Author

@pytorchbot merge

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

iupaikov-amd pushed a commit to ROCm/pytorch that referenced this pull request Jun 4, 2025
For graph partition, `write_get_raw_stream_header_once` is done once so the autotune code may not have the header. This PR additionally calls `write_get_raw_stream_header` in `codegen_device_guard_enter` before `get_raw_stream` is used.

Pull Request resolved: pytorch#154698
Approved by: https://github.com/oulgen
@BoyuanFeng BoyuanFeng mentioned this pull request May 22, 2025
20 tasks
@github-actions github-actions bot deleted the bf/partition-standalone-compile branch July 4, 2025 02:22
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants