[FlexFlash] Specify lowering w/ new BACKEND kernel option#168017
[FlexFlash] Specify lowering w/ new BACKEND kernel option#168017drisspg wants to merge 12 commits intogh/drisspg/221/basefrom
BACKEND kernel option#168017Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/168017
Note: Links to docs will display an error until the docs builds have been completed. ✅ You can merge normally! (2 Unrelated Failures)As of commit f17a038 with merge base 39ebab1 ( BROKEN TRUNK - The following job failed but were present on the merge base:👉 Rebase onto the `viable/strict` branch to avoid these failures
UNSTABLE - The following job is marked as unstable, possibly due to flakiness on trunk:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
ghstack-source-id: d4bfe5a Pull-Request: pytorch/pytorch#168017
BACKEND kernel option
|
@pytorchbot merge |
Merge startedYour change will be merged once all checks pass (ETA 0-4 Hours). Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
| mask_graph: Subgraph, | ||
| kernel_options: dict[str, Any], | ||
| num_score_mod_placeholders: int, | ||
| force_impl: Literal["AUTO", "TRITON", "FLASH", "TRITON_DECODE"], |
There was a problem hiding this comment.
May also want to use the private _Backend literal here imported from the other file is what I meant with mh last comment. force_impl: _BACKEND
There was a problem hiding this comment.
yeah, ill let this land and do it in the top commit
Merge failedReason: 1 jobs have failed, first few of them are: trunk / linux-jammy-cuda12.8-py3.10-gcc11 / test (default, 1, 5, lf.linux.g6.4xlarge.experimental.nvidia.gpu) Details for Dev Infra teamRaised by workflow job |
|
@pytorchbot merge -f "insanely long ci times" |
Merge startedYour change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes). Please use Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
Need to land: Dao-AILab/flash-attention#1985 ^^First^^ Pull Request resolved: #167040 Approved by: https://github.com/Skylion007, https://github.com/albanD ghstack dependencies: #168017
ghstack-source-id: 9319e5b Pull-Request: pytorch/pytorch#168017
Stack from ghstack (oldest at bottom):
Align w/ naming convention
cc @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @chenyang78 @kadeng @muchulee8 @amjames @chauhang @aakhundov @coconutruben @Chillee @yanboliang @BoyuanFeng