Skip to content

New conversations now auto-select the first loaded model#20403

Merged
ServeurpersoCom merged 2 commits intoggml-org:masterfrom
ServeurpersoCom:autoselect-model-new-conv
Mar 12, 2026
Merged

New conversations now auto-select the first loaded model#20403
ServeurpersoCom merged 2 commits intoggml-org:masterfrom
ServeurpersoCom:autoselect-model-new-conv

Conversation

@ServeurpersoCom
Copy link
Contributor

In router mode, new conversations now auto-select the first loaded model
instead of showing an empty dropdown.

Existing conversations still pick up the last model used in each
conversation when switching between chats.

Solves #20382

@ServeurpersoCom ServeurpersoCom merged commit de19015 into ggml-org:master Mar 12, 2026
8 checks passed
tekintian added a commit to tekintian/llama.cpp that referenced this pull request Mar 12, 2026
* 'master' of github.com:ggml-org/llama.cpp: (33 commits)
  convert : better mtp check and fix return [no ci] (ggml-org#20419)
  vulkan: fix SSM_CONV PP scaling with large ubatch sizes (ggml-org#20379)
  New conversations now auto-select the first loaded model (ggml-org#20403)
  ggml-virtgpu: Fix some build commands (ggml-org#20341)
  metal : avoid divisions in bin kernel (ggml-org#20426)
  ci: Setup self-hosted CI for Intel Linux Vulkan backend (ggml-org#20154)
  vulkan: fix l2_norm epsilon handling (ggml-org#20350)
  vulkan: fix OOB check in flash_attn_mask_opt (ggml-org#20296)
  vulkan: Fix ErrorOutOfHostMemory on Intel GPU when loading large models with --no-mmap (ggml-org#20059)
  opencl: use larger workgroup size for get_rows (ggml-org#20316)
  opencl: add cumsum op (ggml-org#18981)
  hip: compile debug builds with -O2 on hip to avoid a compiler bug (ggml-org#20392)
  common/parser: add GigaChatV3/3.1 models support (ggml-org#19931)
  model : add support for Phi4ForCausalLMV (ggml-org#20168)
  graph : add optional scale parameter to build_lora_mm [no ci] (ggml-org#20427)
  common : fix --n-cpu-moe, --cpu-moe for models with fused gate + up (ggml-org#20416)
  ggml-webgpu: Add supports for `GGML_OP_REPEAT` (ggml-org#20230)
  llama : enable chunked fused GDN path (ggml-org#20340)
  llama : whitespace cleanup (ggml-org#20422)
  ggml : add NVFP4 quantization type support (ggml-org#19769)
  ...
am17an pushed a commit to am17an/llama.cpp that referenced this pull request Mar 12, 2026
)

* webui: auto-select first loaded model for new conversations in router mode

* chore: update webui build output
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants