Update FlexAttention with masking semantic#133373
Update FlexAttention with masking semantic#133373drisspg wants to merge 8 commits intogh/drisspg/30/basefrom
Conversation
[ghstack-poisoned]
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/133373
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 5dd6433 with merge base 14ddd93 ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy yf225 chenyang78 kadeng muchulee8 ColinPeppler amjames desertfire chauhang [ghstack-poisoned]
|
@pytorchbot merge |
Merge startedYour change will be merged once all checks pass (ETA 0-4 Hours). Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
|
The merge job was canceled or timed out. This most often happen if two merge requests were issued for the same PR, or if merge job was waiting for more than 6 hours for tests to finish. In later case, please do not hesitate to reissue the merge command |
cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy yf225 chenyang78 kadeng muchulee8 ColinPeppler amjames desertfire chauhang [ghstack-poisoned]
Chillee
left a comment
There was a problem hiding this comment.
Can you add a test like this?
out, lse = torch.compile(flex_attention)(q, k, v, score_mod=mask, return_lse=True)
lse.sum().backward()
cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy yf225 chenyang78 kadeng muchulee8 ColinPeppler amjames desertfire chauhang [ghstack-poisoned]
cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy yf225 chenyang78 kadeng muchulee8 ColinPeppler amjames desertfire chauhang [ghstack-poisoned]
cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy yf225 chenyang78 kadeng muchulee8 ColinPeppler amjames desertfire chauhang [ghstack-poisoned]
Reverting PR 133373 failedReason: Command Details for Dev Infra teamRaised by workflow job |
This reverts commit e847b6b. Reverted #134043 on behalf of https://github.com/jeanschmidt due to Need to revert, in order to be able to revert #133373, feel free to reland this after solving conflicts ([comment](#134043 (comment)))
# Summary Adds the option for the head dims to be different between QK and V tensors. Fixes issue: #133674 V_DIM > QK_DIM is blocked by landing: triton-lang/triton#4138 / triton-lang/triton#4540 Into PyTorch's triton branch. Pull Request resolved: #134043 Approved by: https://github.com/Chillee
|
@pytorchbot revert -m "Broke internal test/inductor signals, see D61611729" -c ghfirst |
|
@pytorchbot successfully started a revert job. Check the current status here. |
This reverts commit 5a7b544. Reverted #133373 on behalf of https://github.com/jeanschmidt due to Broke internal test/inductor signals, see D61611729 ([comment](#133373 (comment)))
|
@drisspg your PR has been successfully reverted. |
cc voznesenskym penguinwu EikanWang jgong5 Guobing-Chen XiaobingSuper zhuhaozhe blzheng wenzhe-nrv jiayisunx ipiszy yf225 chenyang78 kadeng muchulee8 ColinPeppler amjames desertfire chauhang mikaylagawarecki [ghstack-poisoned]
|
@pytorchbot merge |
Merge startedYour change will be merged once all checks pass (ETA 0-4 Hours). Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
|
@pytorchbot merge -f "very slow queue wait time for rocm jobs" |
|
The merge job was canceled or timed out. This most often happen if two merge requests were issued for the same PR, or if merge job was waiting for more than 6 hours for tests to finish. In later case, please do not hesitate to reissue the merge command |
Merge startedYour change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes). Please use Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
…134043)" This reverts commit e847b6b. Reverted pytorch#134043 on behalf of https://github.com/jeanschmidt due to Need to revert, in order to be able to revert pytorch#133373, feel free to reland this after solving conflicts ([comment](pytorch#134043 (comment)))
This reverts commit 5a7b544. Reverted pytorch#133373 on behalf of https://github.com/jeanschmidt due to Broke internal test/inductor signals, see D61611729 ([comment](pytorch#133373 (comment)))
Stack from ghstack (oldest at bottom):
cc @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @yf225 @chenyang78 @kadeng @muchulee8 @ColinPeppler @amjames @desertfire @chauhang @mikaylagawarecki