You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[ROCm] Make bare-host ROCm install self-sufficient
Moves GPU-vendor-specific runtime deps out of common.txt into
requirements/cuda_core.txt and requirements/rocm_core.txt. setup.py
reads common.txt plus whichever core file matches BUILD_WITH_HIP so
`pip install -e .` Just Works on both CUDA and ROCm hosts.
- Drop cupy-cuda12x and nixl from common.txt (both are CUDA-only on PyPI;
the nixl meta-package unconditionally pulls nixl-cu12, which installs
nixl_ep/ and breaks ROCm runtime).
- cuda.txt now -r cuda_core.txt so Dockerfile's `pip install -r cuda.txt`
still pulls the same set.
- Remove the [tool.setuptools.dynamic] dependencies block from
pyproject.toml; install_requires is driven by setup.py now.
- Add a second "Without vLLM docker base image" subsection to the
ROCm install docs, mirroring the CUDA from-source flow line-for-line
(uv venv -> -r build.txt -> torch from ROCm wheel index -> build).
The existing rocm/vllm-dev flow stays as-is.
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
Signed-off-by: Shaoting Feng <shaotingf@uchicago.edu>
The `AMD Infinity hub <https://hub.docker.com/r/rocm/vllm-dev>`__ for vLLM offers a prebuilt, optimized docker image designed for validating inference performance on the AMD Instinct™ MI300X accelerator.
212
215
The image is based on the latest vLLM v1. Please check `LLM inference performance validation on AMD Instinct MI300X <https://rocm.docs.amd.com/en/latest/how-to/rocm-for-ai/inference/benchmark-docker/vllm.html?model=pyt_vllm_llama-3.1-8b>`__ for instructions on how to use this prebuilt docker image.
@@ -235,7 +238,7 @@ As of the date of writing, the steps are validated on the following environment:
235
238
bash
236
239
237
240
Install Latest LMCache from Source for ROCm
238
-
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
241
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
239
242
240
243
To install from source, clone the repository and install in editable mode.
241
244
@@ -255,4 +258,37 @@ Example on MI300X (gfx942):
255
258
TORCH_DONT_CHECK_COMPILER_ABI=1 \
256
259
CXX=hipcc \
257
260
BUILD_WITH_HIP=1 \
258
-
python3 -m pip install --no-build-isolation -e .
261
+
python3 -m pip install --no-build-isolation -e .
262
+
263
+
264
+
On a bare ROCm host
265
+
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
266
+
267
+
Install Latest LMCache from Source for ROCm
268
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
269
+
270
+
To install from source on a bare ROCm host (no ``rocm/vllm-dev`` base image),
271
+
torch must be installed from the ROCm wheel index before building LMCache.
272
+
This mirrors the CUDA from-source flow above, with the ROCm wheel index and
273
+
HIP build flags in place of their CUDA equivalents.
274
+
275
+
.. code-block:: bash
276
+
277
+
git clone https://github.com/LMCache/LMCache.git
278
+
cd LMCache
279
+
280
+
uv venv --python 3.12
281
+
source .venv/bin/activate
282
+
283
+
# Need to install these packages manually to avoid build isolation
0 commit comments