Skip modded_nanogpt model in TorchInductor benchmark#172125
Skip modded_nanogpt model in TorchInductor benchmark#172125huydhn wants to merge 1 commit intopytorch:mainfrom
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/172125
Note: Links to docs will display an error until the docs builds have been completed. ⏳ No Failures, 28 PendingAs of commit 067e3d5 with merge base db4f3aa ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
can you file an issue saying "this has never worked correctly before" and assign it to whoever added the model originally? |
|
@pytorchbot rebase |
|
@pytorchbot started a rebase job onto refs/remotes/origin/viable/strict. Check the current status here |
This has never been working correctly pytorch#172015 (comment) Signed-off-by: Huy Do <huydhn@gmail.com>
|
Successfully rebased |
7890854 to
067e3d5
Compare
|
@pytorchbot drci |
|
@pytorchbot merge -f 'Should be ok to force merge, lint jobs are green' |
Yes, @atalman mentions the same thing, let me create a separate issue for that |
Merge startedYour change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes). Please use Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
|
@pytorchbot cherry-pick --onto release/2.10 --fixes 'Skip modded_nanogpton TorchInductor benchmark' -c regression |
This has never been working correctly #172015 (comment), so let's skip it completely. For more context, this model switches from [eager_fail_to_run to fail_accuracy](https://hud.pytorch.org/benchmark/v3/dashboard/compiler_inductor?renderGroupId=main&time.start=2026-01-02T00%3A00%3A00.000Z&time.end=2026-01-09T23%3A59%3A59.999Z&filters.repo=pytorch%2Fpytorch&filters.benchmarkName=compiler&filters.backend=&filters.mode=training&filters.dtype=amp&filters.deviceName=cuda+%28h100%29&filters.device=cuda&filters.arch=h100&filters.model=modded_nanogpt&lbranch=prepare-perf-number-2.9.1&rbranch=prepare-perf-number-2.10-v2&rcommit.commit=48ac90a47e9eb80c30282ae991f2bb3399427919&rcommit.workflow_id=20837908962&rcommit.date=2026-01-09T04%3A00%3A00Z&rcommit.branch=prepare-perf-number-2.10-v2&lcommit.commit=be32f077fb50b49ad856e957a6860d37565583db&lcommit.workflow_id=20799100249&lcommit.date=2026-01-08T02%3A00%3A00Z&lcommit.branch=prepare-perf-number-2.9.1&maxSampling=110) status recently, root cause is not clear, but it's not related to 2.10 RC Pull Request resolved: #172125 Approved by: https://github.com/zou3519, https://github.com/atalman (cherry picked from commit 0cea083)
Cherry picking #172125The cherry pick PR is at #172141 and it is linked with issue Skip modded_nanogpton TorchInductor benchmark. The following tracker issues are updated: Details for Dev Infra teamRaised by workflow job |
Skip modded_nanogpt model in TorchInductor benchmark (#172125) This has never been working correctly #172015 (comment), so let's skip it completely. For more context, this model switches from [eager_fail_to_run to fail_accuracy](https://hud.pytorch.org/benchmark/v3/dashboard/compiler_inductor?renderGroupId=main&time.start=2026-01-02T00%3A00%3A00.000Z&time.end=2026-01-09T23%3A59%3A59.999Z&filters.repo=pytorch%2Fpytorch&filters.benchmarkName=compiler&filters.backend=&filters.mode=training&filters.dtype=amp&filters.deviceName=cuda+%28h100%29&filters.device=cuda&filters.arch=h100&filters.model=modded_nanogpt&lbranch=prepare-perf-number-2.9.1&rbranch=prepare-perf-number-2.10-v2&rcommit.commit=48ac90a47e9eb80c30282ae991f2bb3399427919&rcommit.workflow_id=20837908962&rcommit.date=2026-01-09T04%3A00%3A00Z&rcommit.branch=prepare-perf-number-2.10-v2&lcommit.commit=be32f077fb50b49ad856e957a6860d37565583db&lcommit.workflow_id=20799100249&lcommit.date=2026-01-08T02%3A00%3A00Z&lcommit.branch=prepare-perf-number-2.9.1&maxSampling=110) status recently, root cause is not clear, but it's not related to 2.10 RC Pull Request resolved: #172125 Approved by: https://github.com/zou3519, https://github.com/atalman (cherry picked from commit 0cea083) Co-authored-by: Huy Do <huydhn@gmail.com>
Summary: This has never been working correctly pytorch/pytorch#172015 (comment), so let's skip it completely. For more context, this model switches from [eager_fail_to_run to fail_accuracy](https://hud.pytorch.org/benchmark/v3/dashboard/compiler_inductor?renderGroupId=main&time.start=2026-01-02T00%3A00%3A00.000Z&time.end=2026-01-09T23%3A59%3A59.999Z&filters.repo=pytorch%2Fpytorch&filters.benchmarkName=compiler&filters.backend=&filters.mode=training&filters.dtype=amp&filters.deviceName=cuda+%28h100%29&filters.device=cuda&filters.arch=h100&filters.model=modded_nanogpt&lbranch=prepare-perf-number-2.9.1&rbranch=prepare-perf-number-2.10-v2&rcommit.commit=48ac90a47e9eb80c30282ae991f2bb3399427919&rcommit.workflow_id=20837908962&rcommit.date=2026-01-09T04%3A00%3A00Z&rcommit.branch=prepare-perf-number-2.10-v2&lcommit.commit=be32f077fb50b49ad856e957a6860d37565583db&lcommit.workflow_id=20799100249&lcommit.date=2026-01-08T02%3A00%3A00Z&lcommit.branch=prepare-perf-number-2.9.1&maxSampling=110) status recently, root cause is not clear, but it's not related to 2.10 RC X-link: pytorch/pytorch#172125 Approved by: https://github.com/zou3519, https://github.com/atalman Reviewed By: atalman, jeanschmidt Differential Revision: D90421963 fbshipit-source-id: 8ae5e397546afd05da131770095f600cb2bea57a
This has never been working correctly pytorch#172015 (comment), so let's skip it completely. For more context, this model switches from [eager_fail_to_run to fail_accuracy](https://hud.pytorch.org/benchmark/v3/dashboard/compiler_inductor?renderGroupId=main&time.start=2026-01-02T00%3A00%3A00.000Z&time.end=2026-01-09T23%3A59%3A59.999Z&filters.repo=pytorch%2Fpytorch&filters.benchmarkName=compiler&filters.backend=&filters.mode=training&filters.dtype=amp&filters.deviceName=cuda+%28h100%29&filters.device=cuda&filters.arch=h100&filters.model=modded_nanogpt&lbranch=prepare-perf-number-2.9.1&rbranch=prepare-perf-number-2.10-v2&rcommit.commit=48ac90a47e9eb80c30282ae991f2bb3399427919&rcommit.workflow_id=20837908962&rcommit.date=2026-01-09T04%3A00%3A00Z&rcommit.branch=prepare-perf-number-2.10-v2&lcommit.commit=be32f077fb50b49ad856e957a6860d37565583db&lcommit.workflow_id=20799100249&lcommit.date=2026-01-08T02%3A00%3A00Z&lcommit.branch=prepare-perf-number-2.9.1&maxSampling=110) status recently, root cause is not clear, but it's not related to 2.10 RC Pull Request resolved: pytorch#172125 Approved by: https://github.com/zou3519, https://github.com/atalman
This has never been working correctly pytorch#172015 (comment), so let's skip it completely. For more context, this model switches from [eager_fail_to_run to fail_accuracy](https://hud.pytorch.org/benchmark/v3/dashboard/compiler_inductor?renderGroupId=main&time.start=2026-01-02T00%3A00%3A00.000Z&time.end=2026-01-09T23%3A59%3A59.999Z&filters.repo=pytorch%2Fpytorch&filters.benchmarkName=compiler&filters.backend=&filters.mode=training&filters.dtype=amp&filters.deviceName=cuda+%28h100%29&filters.device=cuda&filters.arch=h100&filters.model=modded_nanogpt&lbranch=prepare-perf-number-2.9.1&rbranch=prepare-perf-number-2.10-v2&rcommit.commit=48ac90a47e9eb80c30282ae991f2bb3399427919&rcommit.workflow_id=20837908962&rcommit.date=2026-01-09T04%3A00%3A00Z&rcommit.branch=prepare-perf-number-2.10-v2&lcommit.commit=be32f077fb50b49ad856e957a6860d37565583db&lcommit.workflow_id=20799100249&lcommit.date=2026-01-08T02%3A00%3A00Z&lcommit.branch=prepare-perf-number-2.9.1&maxSampling=110) status recently, root cause is not clear, but it's not related to 2.10 RC Pull Request resolved: pytorch#172125 Approved by: https://github.com/zou3519, https://github.com/atalman
- Re-enable `detectron2_maskrcnn` skip in skip.all. - Re-enable all `timm_*` model skips in skip.all. - Keep explicit upstream PR context comments for `modded_nanogpt` and `pytorch_CycleGAN_and_pix2pix`. - Remove stale expected-accuracy rows for skipped models. Relevant PRs: [1] pytorch/pytorch#120299 [2] pytorch/pytorch#164816 [3] pytorch/pytorch#172125 [4] pytorch/pytorch#175066 [5] #2306
This has never been working correctly #172015 (comment), so let's skip it completely. For more context, this model switches from eager_fail_to_run to fail_accuracy status recently, root cause is not clear, but it's not related to 2.10 RC
cc @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @kadeng @chauhang @amjames @Lucaskabela @jataylo