Checklist
Describe the bug
on latest image it seems --served-model-name gpt-oss:120b has some issues, but after replace : with - like gpt-oss-120b will works.
error message as follows:
openai.BadRequestError: Error code: 400 - {'object': 'error', 'message': 'The following requested LoRA adapters are not loaded: 120b\nLoaded adapters: dict_keys([]).', 'type': 'BadRequestError', 'param': None, 'code': 400}
BadRequestError("Error code: 400 - {'object': 'error', 'message': 'The following requested LoRA adapters are not loaded: 120b\nLoaded adapters: dict_keys([]).', 'type': 'BadRequestError', 'param': None, 'code': 400}")Traceback (most recent call last):
File ".venv/lib/python3.12/site-packages/langchain_core/runnables/base.py", line 3047, in invoke
input_ = context.run(step.invoke, input_, config)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 372, in invoke
self.generate_prompt(
File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 957, in generate_prompt
return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 776, in generate
self._generate_with_cache(
File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 1022, in _generate_with_cache
result = self._generate(
^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/langchain_openai/chat_models/base.py", line 995, in _generate
response = self.client.create(**payload)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/openai/_utils/_utils.py", line 287, in wrapper
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/openai/resources/chat/completions/completions.py", line 925, in create
return self._post(
^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/openai/_base_client.py", line 1242, in post
return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/openai/_base_client.py", line 1037, in request
raise self._make_status_error_from_response(err.response) from None
Reproduction
CLI with issues:
python3 -m sglang.launch_server --model openai/gpt-oss-120b --tp 2 --port 18000 --host 0.0.0.0 --tool-call-parser gpt-oss --reasoning-parser gpt-oss --served-model-name gpt-oss:120b --attention-backend triton
CLI works.
python3 -m sglang.launch_server --model openai/gpt-oss-120b --tp 2 --port 18000 --host 0.0.0.0 --tool-call-parser gpt-oss --reasoning-parser gpt-oss --served-model-name gpt-oss-120b --attention-backend triton
Environment
images info as follows:
lmsysorg/sglang latest 496b1acced6
do not saw same issue on old images which I pulled.
old images info:
lmsysorg/sglang bb0aa0832b2b
Checklist
Describe the bug
on latest image it seems --served-model-name gpt-oss:120b has some issues, but after replace : with - like gpt-oss-120b will works.
error message as follows:
openai.BadRequestError: Error code: 400 - {'object': 'error', 'message': 'The following requested LoRA adapters are not loaded: 120b\nLoaded adapters: dict_keys([]).', 'type': 'BadRequestError', 'param': None, 'code': 400}
BadRequestError("Error code: 400 - {'object': 'error', 'message': 'The following requested LoRA adapters are not loaded: 120b\nLoaded adapters: dict_keys([]).', 'type': 'BadRequestError', 'param': None, 'code': 400}")Traceback (most recent call last):
File ".venv/lib/python3.12/site-packages/langchain_core/runnables/base.py", line 3047, in invoke
input_ = context.run(step.invoke, input_, config)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 372, in invoke
self.generate_prompt(
File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 957, in generate_prompt
return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 776, in generate
self._generate_with_cache(
File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 1022, in _generate_with_cache
result = self._generate(
^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/langchain_openai/chat_models/base.py", line 995, in _generate
response = self.client.create(**payload)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/openai/_utils/_utils.py", line 287, in wrapper
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/openai/resources/chat/completions/completions.py", line 925, in create
return self._post(
^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/openai/_base_client.py", line 1242, in post
return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File ".venv/lib/python3.12/site-packages/openai/_base_client.py", line 1037, in request
raise self._make_status_error_from_response(err.response) from None
Reproduction
CLI with issues:
python3 -m sglang.launch_server --model openai/gpt-oss-120b --tp 2 --port 18000 --host 0.0.0.0 --tool-call-parser gpt-oss --reasoning-parser gpt-oss --served-model-name gpt-oss:120b --attention-backend triton
CLI works.
python3 -m sglang.launch_server --model openai/gpt-oss-120b --tp 2 --port 18000 --host 0.0.0.0 --tool-call-parser gpt-oss --reasoning-parser gpt-oss --served-model-name gpt-oss-120b --attention-backend triton
Environment
images info as follows:
lmsysorg/sglang latest 496b1acced6
do not saw same issue on old images which I pulled.
old images info:
lmsysorg/sglang bb0aa0832b2b