Skip to content

[xpu][feature] Supports woq_int8 inductor pattern on Intel GPU#163615

Closed
xiaowangintel wants to merge 3 commits intopytorch:mainfrom
xiaowangintel:xw/woq_int8_inductor
Closed

[xpu][feature] Supports woq_int8 inductor pattern on Intel GPU#163615
xiaowangintel wants to merge 3 commits intopytorch:mainfrom
xiaowangintel:xw/woq_int8_inductor

Conversation

@xiaowangintel
Copy link
Contributor

@xiaowangintel xiaowangintel commented Sep 23, 2025

Summary:

Supports woq_int8 inductor pattern on Intel GPU. When using torch.compile, woq_int8 will be lowering to _weight_int8pack_mm instead of being falled back mul().sum(). The Intel GPU backend of _weight_int8pack_mm was supported in #160938.

cc @jgong5 @mingfeima @XiaobingSuper @sanchitintel @ashokei @jingxu10 @voznesenskym @penguinwu @EikanWang @Guobing-Chen @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @kadeng @muchulee8 @amjames @chauhang @aakhundov @coconutruben @jataylo @chenyang78

@pytorch-bot
Copy link

pytorch-bot bot commented Sep 23, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/163615

Note: Links to docs will display an error until the docs builds have been completed.

✅ You can merge normally! (1 Unrelated Failure)

As of commit 4c3d0c5 with merge base 0449e64 (image):

FLAKY - The following job failed but was likely due to flakiness present on trunk:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

def matcher_check_fn():
self.assertEqual(
counters["inductor"]["woq_matcher_count"], 0 if TEST_ACL else 1
devices = ["cpu", "xpu"]
Copy link
Collaborator

@etaf etaf Sep 23, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hi, this will fail on device without xpu.
I suggest you move this case to TestPatternMatcherGeneric, which is already designed to be reused by both cpu and xpu.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Gets it, I have modified.

@jbschlosser jbschlosser added triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module intel This tag is for PR from Intel labels Sep 24, 2025
@pytorch-bot pytorch-bot bot removed the ciflow/xpu Run XPU CI tasks label Sep 29, 2025
@etaf etaf added the ciflow/xpu Run XPU CI tasks label Oct 14, 2025
@pytorch-bot
Copy link

pytorch-bot bot commented Oct 14, 2025

To add the ciflow label ciflow/xpu please first approve the workflows that are awaiting approval (scroll to the bottom of this page).

This helps ensure we don't trigger CI on this PR until it is actually authorized to do so. Please ping one of the reviewers if you do not have access to approve and run workflows.

@pytorch-bot pytorch-bot bot removed the ciflow/xpu Run XPU CI tasks label Oct 14, 2025
@etaf etaf added the ciflow/xpu Run XPU CI tasks label Oct 14, 2025
@etaf etaf requested a review from EikanWang October 14, 2025 01:05
@pytorch-bot pytorch-bot bot removed the ciflow/xpu Run XPU CI tasks label Oct 15, 2025
@etaf etaf added the ciflow/xpu Run XPU CI tasks label Oct 15, 2025
@etaf
Copy link
Collaborator

etaf commented Oct 16, 2025

@pytorchbot rebase -b main

@pytorchmergebot
Copy link
Collaborator

@pytorchbot started a rebase job onto refs/remotes/origin/main. Check the current status here

@pytorchmergebot
Copy link
Collaborator

Successfully rebased xw/woq_int8_inductor onto refs/remotes/origin/main, please pull locally before adding more changes (for example, via git checkout xw/woq_int8_inductor && git pull --rebase)

@pytorch-bot pytorch-bot bot removed the ciflow/xpu Run XPU CI tasks label Oct 16, 2025
@etaf etaf added the ciflow/xpu Run XPU CI tasks label Oct 16, 2025
@etaf
Copy link
Collaborator

etaf commented Oct 20, 2025

@pytorchbot rebase -b main

@pytorchmergebot
Copy link
Collaborator

@pytorchbot started a rebase job onto refs/remotes/origin/main. Check the current status here

@pytorchmergebot
Copy link
Collaborator

Successfully rebased xw/woq_int8_inductor onto refs/remotes/origin/main, please pull locally before adding more changes (for example, via git checkout xw/woq_int8_inductor && git pull --rebase)

@pytorch-bot pytorch-bot bot removed the ciflow/xpu Run XPU CI tasks label Oct 20, 2025
@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@pytorchmergebot
Copy link
Collaborator

@pytorch-bot pytorch-bot bot removed ciflow/trunk Trigger trunk jobs on your pull request ciflow/xpu Run XPU CI tasks labels Dec 16, 2025
@xiaowangintel
Copy link
Contributor Author

@pytorchbot merge

@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Dec 16, 2025
@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@pytorchmergebot
Copy link
Collaborator

Merge failed

Reason: 1 jobs have failed, first few of them are: linux-aarch64 / linux-jammy-aarch64-py3.10 / test (default, 3, 3, linux.arm64.m8g.4xlarge)

Details for Dev Infra team Raised by workflow job

@pytorch-bot pytorch-bot bot removed the ciflow/trunk Trigger trunk jobs on your pull request label Dec 18, 2025
@chuanqi129 chuanqi129 added the ciflow/xpu Run XPU CI tasks label Dec 18, 2025
@xiaowangintel
Copy link
Contributor Author

@pytorchbot merge

@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Dec 19, 2025
@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

xgz2 pushed a commit that referenced this pull request Dec 22, 2025
Summary:

Supports woq_int8 inductor pattern on Intel GPU. When using torch.compile, woq_int8 will be lowering to _weight_int8pack_mm instead of being falled back mul().sum(). The Intel GPU backend of _weight_int8pack_mm was supported in #160938.

Pull Request resolved: #163615
Approved by: https://github.com/etaf, https://github.com/EikanWang, https://github.com/desertfire, https://github.com/jansel
krastogi-in pushed a commit to krastogi-in/pytorch that referenced this pull request Jan 9, 2026
…ch#163615)

Summary:

Supports woq_int8 inductor pattern on Intel GPU. When using torch.compile, woq_int8 will be lowering to _weight_int8pack_mm instead of being falled back mul().sum(). The Intel GPU backend of _weight_int8pack_mm was supported in pytorch#160938.

Pull Request resolved: pytorch#163615
Approved by: https://github.com/etaf, https://github.com/EikanWang, https://github.com/desertfire, https://github.com/jansel
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ciflow/trunk Trigger trunk jobs on your pull request ciflow/xpu Run XPU CI tasks intel This tag is for PR from Intel Merged module: inductor open source topic: not user facing topic category triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module

Projects

None yet

Development

Successfully merging this pull request may close these issues.

9 participants