Skip to content

[Bug] The following requested LoRA adapters are not loaded: 120b #12745

@sloan-wu

Description

@sloan-wu

Checklist

  • 1. I have searched related issues but cannot get the expected help.
  • 2. The bug has not been fixed in the latest version.
  • 3. Please note that if the bug-related issue you submitted lacks corresponding environment info and a minimal reproducible demo, it will be challenging for us to reproduce and resolve the issue, reducing the likelihood of receiving feedback.
  • 4. If the issue you raised is not a bug but a question, please raise a discussion at https://github.com/sgl-project/sglang/discussions/new/choose Otherwise, it will be closed.
  • 5. Please use English, otherwise it will be closed.

Describe the bug

on latest image it seems --served-model-name gpt-oss:120b has some issues, but after replace : with - like gpt-oss-120b will works.

error message as follows:
openai.BadRequestError: Error code: 400 - {'object': 'error', 'message': 'The following requested LoRA adapters are not loaded: 120b\nLoaded adapters: dict_keys([]).', 'type': 'BadRequestError', 'param': None, 'code': 400}

BadRequestError("Error code: 400 - {'object': 'error', 'message': 'The following requested LoRA adapters are not loaded: 120b\nLoaded adapters: dict_keys([]).', 'type': 'BadRequestError', 'param': None, 'code': 400}")Traceback (most recent call last):

File ".venv/lib/python3.12/site-packages/langchain_core/runnables/base.py", line 3047, in invoke
input_ = context.run(step.invoke, input_, config)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 372, in invoke
self.generate_prompt(

File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 957, in generate_prompt
return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 776, in generate
self._generate_with_cache(

File ".venv/lib/python3.12/site-packages/langchain_core/language_models/chat_models.py", line 1022, in _generate_with_cache
result = self._generate(
^^^^^^^^^^^^^^^

File ".venv/lib/python3.12/site-packages/langchain_openai/chat_models/base.py", line 995, in _generate
response = self.client.create(**payload)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File ".venv/lib/python3.12/site-packages/openai/_utils/_utils.py", line 287, in wrapper
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^

File ".venv/lib/python3.12/site-packages/openai/resources/chat/completions/completions.py", line 925, in create
return self._post(
^^^^^^^^^^^

File ".venv/lib/python3.12/site-packages/openai/_base_client.py", line 1242, in post
return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File ".venv/lib/python3.12/site-packages/openai/_base_client.py", line 1037, in request
raise self._make_status_error_from_response(err.response) from None

Reproduction

CLI with issues:
python3 -m sglang.launch_server --model openai/gpt-oss-120b --tp 2 --port 18000 --host 0.0.0.0 --tool-call-parser gpt-oss --reasoning-parser gpt-oss --served-model-name gpt-oss:120b --attention-backend triton

CLI works.
python3 -m sglang.launch_server --model openai/gpt-oss-120b --tp 2 --port 18000 --host 0.0.0.0 --tool-call-parser gpt-oss --reasoning-parser gpt-oss --served-model-name gpt-oss-120b --attention-backend triton

Environment

images info as follows:
lmsysorg/sglang latest 496b1acced6

do not saw same issue on old images which I pulled.
old images info:
lmsysorg/sglang bb0aa0832b2b

Metadata

Metadata

Assignees

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions