-
Notifications
You must be signed in to change notification settings - Fork 7.4k
Deepseek OCR unsupported #58702
Copy link
Copy link
Closed
Labels
bugSomething that is supposed to be working; but isn'tSomething that is supposed to be working; but isn'tcommunity-backlogllmquestionJust a question :)Just a question :)serveRay Serve Related IssueRay Serve Related Issuestability
Description
What happened + What you expected to happen
[<Ray component: Core|RLlib|etc...>] deepseek OCR isnt supported. the deployment of VLLM nightly that
TypeError: init_app_state() got an unexpected keyword argument 'vllm_config'
Versions / Dependencies
vllm/vllm-openai:nightly-f849ee739cdb3d82fce1660a6fd91806e8ae9bff
ray 3.0.0.dev0
Reproduction script
serve deploy config.yaml
with the config file being the following
host: 0.0.0.0
port: 8001
# takes about 1:30 minutes to change the cluster to RUNNING mode on EC2 g5.2xlarge
applications:
- import_path: ray.serve.llm:build_openai_app
name: Multimodal LLMs
route_prefix: "/"
args:
llm_configs:
- model_loading_config:
model_id: deepseek-ai/DeepSeek-OCR
model_source: deepseek-ai/DeepSeek-OCR # Qwen/Qwen2.5-VL-7B-Instruct
engine_kwargs:
max_model_len: 8192 # 12288 # safe for 24GB VRAM
# dtype: bfloat16
gpu_memory_utilization: 0.9
ray_actor_options:
resources: {"gpu": 1} # Use default GPU resource```
### Issue Severity
None
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
bugSomething that is supposed to be working; but isn'tSomething that is supposed to be working; but isn'tcommunity-backlogllmquestionJust a question :)Just a question :)serveRay Serve Related IssueRay Serve Related Issuestability