Add shell script to deploy LLM service on AMD GPU#2321
Add shell script to deploy LLM service on AMD GPU#2321cb-github-robot merged 4 commits intocloud-barista:mainfrom
Conversation
|
@leehyeoklee Let's check if the suggested scripts can be (simply) merged with the existing scripts. :)
https://github.com/cloud-barista/cb-tumblebug/tree/main/scripts/usecases/llm |
|
@leehyeoklee |
66ef178 to
5742ef7
Compare
|
Yes, it's ready for another round of review.😊 Additionally, I’ve created a single installGpuDriver.sh script that handles both GPU driver and CUDA/ROCm installations. And I've confirmed that LLM models load and run correctly on both NVIDIA and AMD GPU VMs! |
|
For AMD vLLM deployment, I referred to this documentation: https://docs.vllm.ai/en/stable/getting_started/installation/gpu/ Note: |
|
/approve |
🚀 Key Changes
This update extends LLM inference and serving capabilities to AMD GPUs, in addition to the existing NVIDIA support.
We have added shell scripts to automate ROCm driver installation, vLLM/Ollama environment setup, and model serving(vLLM). This allows users on cloud environments (Azure, AWS) with AMD GPUs to deploy LLM services without complex manual configuration.
✨ Implementation Details
1. ROCm Driver Installation (
installRocmDriver.sh)2. vLLM Environment Deployment (
deployvLLMAmd.sh)rocm/vllmDocker image provided by AMD.3. vLLM Model Serving (
servevLLMAmd.sh)💻 Test Environment and Results
✅ Azure (Radeon PRO V710)
Ollama

vLLM
