[CI] Disable TRANSFORMERS_OFFLINE for nightly vLLM benchmark runs#176553
[CI] Disable TRANSFORMERS_OFFLINE for nightly vLLM benchmark runs#176553huydhn wants to merge 1 commit intopytorch:mainfrom
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/176553
Note: Links to docs will display an error until the docs builds have been completed. ⏳ 25 Pending, 1 Unrelated FailureAs of commit 067833d with merge base 9dfac50 ( FLAKY - The following job failed but was likely due to flakiness present on trunk:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
|
Nightly scheduled runs should be able to download from HuggingFace directly rather than being restricted to the local cache. This adds an `is_nightly` input to the reusable workflow, set to true when the caller detects `github.event_name == 'schedule'`. Manual dispatch runs continue to use the cache (TRANSFORMERS_OFFLINE=1). Includes a comment explaining the rationale for the nightly override. Signed-off-by: Huy Do <huydhn@gmail.com>
789b1dc to
067833d
Compare
|
@pytorchbot merge -f 'CI only tweak, should be fine' |
Merge startedYour change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes). Please use Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
…torch#176553) Instead of always using offline mode, we need to have a way to periodically refresh the local cache. This job runs only once per day, so rate limit wouldn't be an issue I observe [a failure in trunk](https://github.com/pytorch/pytorch/actions/runs/22625578355/job/65575251092#step:15:13837) for `pytorch/gemma-3-12b-it-int4` where the local cache doesn't seem to work anymore, not sure why, could be due to the recent transformers version update. The failure goes away when I turn off TRANSFORMERS_OFFLINE to refresh the local cache. Pull Request resolved: pytorch#176553 Approved by: https://github.com/zou3519
Instead of always using offline mode, we need to have a way to periodically refresh the local cache. This job runs only once per day, so rate limit wouldn't be an issue
I observe a failure in trunk for
pytorch/gemma-3-12b-it-int4where the local cache doesn't seem to work anymore, not sure why, could be due to the recent transformers version update. The failure goes away when I turn off TRANSFORMERS_OFFLINE to refresh the local cache.