Skip to content

[CI] Split V1 Others into 3 separate jobs#37016

Merged
khluu merged 1 commit intovllm-project:mainfrom
khluu:ci/split-v1-others
Mar 23, 2026
Merged

[CI] Split V1 Others into 3 separate jobs#37016
khluu merged 1 commit intovllm-project:mainfrom
khluu:ci/split-v1-others

Conversation

@khluu
Copy link
Copy Markdown
Collaborator

@khluu khluu commented Mar 13, 2026

Summary

  • Split the single ~55m "V1 Others" job into three smaller jobs targeting ~20m each
  • V1 Spec Decode (~19m): spec_decode tests (the heaviest single test suite)
  • V1 Sample + Logits (~18m): sample (11m), logits_processors (6m), test_oracle, test_request, test_outputs
  • V1 Core + KV + Metrics (~18m): core (1m), executor (1m), kv_offload (3m), worker (1m), kv_connector/unit (6m), metrics (1m), lm_eval integration (5m)
  • Each job has properly scoped source_file_dependencies instead of the broad tests/v1
  • AMD mirrors preserved for all three jobs

Test plan

  • Verify all three new jobs pass in CI
  • Confirm no tests are missing compared to the original single job
  • Check that source_file_dependencies correctly trigger each job

AI assistance was used (Claude). This is not duplicating any existing PR.

🤖 Generated with Claude Code

Split the single ~55m "V1 Others" job into three smaller jobs
targeting ~20m each:

- V1 Spec Decode (~19m): spec_decode tests
- V1 Sample + Logits (~18m): sample, logits_processors, test_oracle,
  test_request, test_outputs
- V1 Core + KV + Metrics (~18m): core, executor, kv_offload, worker,
  kv_connector/unit, metrics, lm_eval integration test

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
@mergify mergify bot added the ci/build label Mar 13, 2026
Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request refactors a single large CI job into three smaller, parallel jobs to improve runtime. The split of tests and dependencies appears logical. However, I've identified a missing configuration in one of the new jobs which contradicts the stated goal in the pull request description and could lead to incomplete test coverage on AMD hardware.

Comment on lines +42 to 65
- label: V1 Core + KV + Metrics
timeout_in_minutes: 30
source_file_dependencies:
- vllm/
- tests/v1/core
- tests/v1/executor
- tests/v1/kv_offload
- tests/v1/worker
- tests/v1/kv_connector/unit
- tests/v1/metrics
- tests/entrypoints/openai/correctness/test_lmeval.py
commands:
- uv pip install --system -r /vllm-workspace/requirements/kv_connectors.txt
- export VLLM_WORKER_MULTIPROC_METHOD=spawn
# split the test to avoid interference
- pytest -v -s -m 'not cpu_test' v1/core
- pytest -v -s v1/executor
- pytest -v -s v1/kv_offload
- pytest -v -s v1/sample
- pytest -v -s v1/logits_processors
- pytest -v -s v1/worker
# TODO: create another `optional` test group for slow tests
- pytest -v -s -m 'not slow_test' v1/spec_decode
- pytest -v -s -m 'not cpu_test' v1/kv_connector/unit
- pytest -v -s -m 'not cpu_test' v1/metrics
- pytest -v -s v1/test_oracle.py
- pytest -v -s v1/test_request.py
- pytest -v -s v1/test_outputs.py
# Integration test for streaming correctness (requires special branch).
- pip install -U git+https://github.com/robertgshaw2-redhat/lm-evaluation-harness.git@streaming-api
- pytest -v -s entrypoints/openai/correctness/test_lmeval.py::test_lm_eval_accuracy_v1_engine
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

The V1 Core + KV + Metrics job is missing the mirror configuration for AMD GPUs. The PR description states that 'AMD mirrors preserved for all three jobs', but this job lacks the configuration present in the other two new jobs. This omission will prevent this set of tests from running on the specified AMD hardware.

- label: V1 Core + KV + Metrics
  timeout_in_minutes: 30
  source_file_dependencies:
    - vllm/
    - tests/v1/core
    - tests/v1/executor
    - tests/v1/kv_offload
    - tests/v1/worker
    - tests/v1/kv_connector/unit
    - tests/v1/metrics
    - tests/entrypoints/openai/correctness/test_lmeval.py
  commands:
    - uv pip install --system -r /vllm-workspace/requirements/kv_connectors.txt
    - export VLLM_WORKER_MULTIPROC_METHOD=spawn
    # split the test to avoid interference
    - pytest -v -s -m 'not cpu_test' v1/core
    - pytest -v -s v1/executor
    - pytest -v -s v1/kv_offload
    - pytest -v -s v1/worker
    - pytest -v -s -m 'not cpu_test' v1/kv_connector/unit
    - pytest -v -s -m 'not cpu_test' v1/metrics
    # Integration test for streaming correctness (requires special branch).
    - pip install -U git+https://github.com/robertgshaw2-redhat/lm-evaluation-harness.git@streaming-api
    - pytest -v -s entrypoints/openai/correctness/test_lmeval.py::test_lm_eval_accuracy_v1_engine
  mirror:
    amd:
      device: mi325_1
      depends_on:
      - image-build-amd

@khluu khluu changed the title [CI] Split V1 Others into 3 parallel jobs [CI] Split V1 Others into 3 separate jobs Mar 13, 2026
@khluu khluu added the ready ONLY add when PR is ready to merge/full CI is needed label Mar 13, 2026
@khluu
Copy link
Copy Markdown
Collaborator Author

khluu commented Mar 14, 2026

V1 Spec Decode : 24m
V1 sample + logits: 24m
V1 core + kv + metrics: 24m
Total 72 minutes, 15 mins more than the original job due to overhead

@AndreasKaratzas
Copy link
Copy Markdown
Collaborator

@khluu is this planned to get merged? Only asking so that I know if I should revert test amd yaml as well.

@khluu
Copy link
Copy Markdown
Collaborator Author

khluu commented Mar 23, 2026

@khluu is this planned to get merged? Only asking so that I know if I should revert test amd yaml as well.

I'll merge it, but for future reference I think you should only change test-amd.yaml once it's confirmed that the changes on test_areas/ get merged

@khluu khluu enabled auto-merge (squash) March 23, 2026 20:54
@khluu khluu merged commit 2488a82 into vllm-project:main Mar 23, 2026
26 checks passed
RhizoNymph pushed a commit to RhizoNymph/vllm that referenced this pull request Mar 26, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
HenryTangDev pushed a commit to HenryTangMain/vllm that referenced this pull request Mar 27, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
khairulkabir1661 pushed a commit to khairulkabir1661/vllm that referenced this pull request Mar 27, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
Monishver11 pushed a commit to Monishver11/vllm that referenced this pull request Mar 27, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
Signed-off-by: Monishver Chandrasekaran <monishverchandrasekaran@gmail.com>
nithinvc pushed a commit to nithinvc/vllm that referenced this pull request Mar 27, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>

Signed-off-by: Nithin Chalapathi <nithin.ch10@gmail.com>
JiantaoXu pushed a commit to JiantaoXu/vllm that referenced this pull request Mar 28, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
vrdn-23 pushed a commit to vrdn-23/vllm that referenced this pull request Mar 30, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
Signed-off-by: Vinay Damodaran <vrdn@hey.com>
EricccYang pushed a commit to EricccYang/vllm that referenced this pull request Apr 1, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
Signed-off-by: EricccYang <yangyang4991@gmail.com>
liuchenbing2026 pushed a commit to liuchenbing2026/vllm that referenced this pull request Apr 4, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
rishitdholakia13 pushed a commit to rishitdholakia13/vllm that referenced this pull request Apr 7, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
Signed-off-by: rishitdholakia13 <rishit+github@cohere.com>
puririshi98 pushed a commit to puririshi98/vllm that referenced this pull request Apr 7, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
Signed-off-by: Rishi Puri <riship@nvidia.com>
big-yellow-duck pushed a commit to EmbeddedLLM/vllm that referenced this pull request Apr 8, 2026
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci/build ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants