Relax flaky B200 GSM8K accuracy thresholds#20304
Merged
Conversation
- DeepSeek V3 FP4: 0.935 → 0.93 (observed values 0.9318–0.9507 across 12 runs) - Eagle DP Attention: 0.64 → 0.62 (observed values 0.630–0.655 across 12 runs)
Contributor
|
Warning You have reached your daily quota limit. Please wait up to 24 hours and I will start processing your requests again! |
hnyls2002
requested changes
Mar 11, 2026
Collaborator
There was a problem hiding this comment.
Remove this file. You can only put it in the Pr's body, not in the repo
hnyls2002
approved these changes
Mar 11, 2026
liubiyongge
pushed a commit
to liubiyongge/sglang
that referenced
this pull request
Mar 13, 2026
Co-authored-by: Alison Shao <alisonshao@Mac.attlocal.net>
Wangzheee
pushed a commit
to Wangzheee/sglang
that referenced
this pull request
Mar 21, 2026
Co-authored-by: Alison Shao <alisonshao@Mac.attlocal.net>
JustinTong0323
pushed a commit
to JustinTong0323/sglang
that referenced
this pull request
Apr 7, 2026
Co-authored-by: Alison Shao <alisonshao@Mac.attlocal.net>
yhyang201
pushed a commit
to yhyang201/sglang
that referenced
this pull request
Apr 22, 2026
Co-authored-by: Alison Shao <alisonshao@Mac.attlocal.net>
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Summary
Both tests run in
stage-c-test-4-gpu-b200and have been causing flaky CI failures due to thresholds set too close to the natural variance of accuracy on a 200-question benchmark.Data Analysis (~1 month of scheduled CI)
Analyzed 200 scheduled CI runs (Jan 20 – Mar 11). Extracted accuracy values from B200 shard logs across
stage-c-test-4-gpu-b200 (0)and(2).Overview
DeepSeek V3 FP4 GSM8K (120 runs, 486 values, Feb 6 – Mar 11)
Sample failures (click run IDs for logs):
Eagle DP Attention GSM8K (93 data points, 103 runs, Feb 10 – Mar 11)
Tracked across shards — the test ran in shard 2 before ~Mar 6 and moved to shard 0 after.
assertGreatermeans== 0.64also fails)Runs with failures (accuracy ≤ 0.64):
Test plan