Skip to content

Fix llm_arch_is_hybrid#1305

Merged
ikawrakow merged 1 commit intomainfrom
ik/fix_hybrid_detection
Feb 23, 2026
Merged

Fix llm_arch_is_hybrid#1305
ikawrakow merged 1 commit intomainfrom
ik/fix_hybrid_detection

Conversation

@ikawrakow
Copy link
Owner

@ikawrakow ikawrakow commented Feb 23, 2026

Fixes hybrid model detection. There was a typo (QWEN3MOE instead of QWEN35MOE).

Thanks to @MrHills-rs for noticing.

@ikawrakow ikawrakow merged commit 2bb40f8 into main Feb 23, 2026
@magikRUKKOLA
Copy link

magikRUKKOLA commented Feb 23, 2026

Any idea why I am unable to run the IQ2_KL at the machine with 256GB RAM?
Woops. I meant the ubergarm/GLM-5-IQ2_KL.

llama_init_from_model: KV self size  = 7458.75 MiB, c^KV (q8_0): 7458.75 MiB, kv^T: not used
llama_init_from_model:  CUDA_Host  output buffer size =     0.59 MiB
ggml_new_object: not enough space in the context's memory pool (needed 5989888, available 5989520)

@MrHills-rs
Copy link

MrHills-rs commented Feb 23, 2026

Any idea why I am unable to run the IQ2_KL at the machine with 256GB RAM?

llama_init_from_model: KV self size  = 7458.75 MiB, c^KV (q8_0): 7458.75 MiB, kv^T: not used
llama_init_from_model:  CUDA_Host  output buffer size =     0.59 MiB
ggml_new_object: not enough space in the context's memory pool (needed 5989888, available 5989520)

I'm running qwen3.5 IQ2_XS on 128gb.
Isn't that the same error I saw when -ub was too high btw? Can you try lowering it to 4096?

Edit: or actually even lower, just to test, say 512

@ikawrakow
Copy link
Owner Author

Any idea why I am unable to run the IQ2_KL at the machine with 256GB RAM?

This is on the latest main branch?

@magikRUKKOLA
Copy link

Any idea why I am unable to run the IQ2_KL at the machine with 256GB RAM?

This is on the latest main branch?

Woops. I confused the LLMs. I meant the ubergarm/GLM-5-IQ2_KL . Interestingly, I can't run it anymore after the recent update.

@ikawrakow
Copy link
Owner Author

@magikRUKKOLA

Does #1306 fix it?

@magikRUKKOLA
Copy link

@ikawrakow

Does #1306 fix it?

Yes, it does.

abc-nix pushed a commit to abc-nix/ik_llama.cpp that referenced this pull request Feb 26, 2026
abc-nix pushed a commit to abc-nix/ik_llama.cpp that referenced this pull request Feb 26, 2026
* Better estimate for max. nuber of compute nodes

* Just in case

server: fix crash from adaptive p (ikawrakow#1304)

Co-authored-by: firecoperana <firecoperana>

Fix tool call for Qwen3.5 (ikawrakow#1300)

* Fix tool call for Qwen3.5

Loosely based on mainline changes from:
* ggml-org/llama.cpp#19635
* ggml-org/llama.cpp#19765

Also need to change the grammar to allow the model to make multiple
tool calls in a row. This was likely broken for Qwen3 Coder prior to
this commit.

* Fix the grammar for the subsequent parameters after the first one

Graph parallel for Qwen3-Next (ikawrakow#1292)

* WIP

* This works, but is slower than split mode layer

Fix llm_arch_is_hybrid (ikawrakow#1305)

Fix max nodes (again) (ikawrakow#1306)

Fix typo in merge-up-gate-experts argument (ikawrakow#1311)

llama-quantize: --dry-run option (ikawrakow#1309)

Slightly better graph parallel for Qwen3-Next (ikawrakow#1307)

* Make sure we pick the reduced tensor from the right GPU

* Minor

Minor delta-net tweak (ikawrakow#1308)

* Make sure we pick the reduced tensor from the right GPU

* Minor

* Minor delta-net tweak

adaptive p: collect probability before logit bias (ikawrakow#1314)

server: propagate task index to response objects for batch requests (ikawrakow#1303)

When multiple prompts are sent in a single /v1/completions request,
each response needs to carry the correct index so the client can
match results to their corresponding prompts. The index field was
not being set on partial responses, final responses, or embedding
responses, causing batch results to all report index 0.

Set res->index = slot.task->index in send_partial_response,
send_final_response, and send_embedding.

Generated with [Devin](https://cli.devin.ai/docs)

Co-authored-by: Joshua Jolley <jjolley@clearwateranalytics.com>
Co-authored-by: Devin <noreply@cognition.ai>

Llama-quantize: Partial requant feature (ikawrakow#1313)

* Partial Requant feature for llama-quantize

- Inspired by the recently portcopied --dry-run feature.
- Allows to partially requantize a split quantized .gguf by requantizing only the missing splits in the destination directory.
- Works both for GGUF which are split tensors by tensors, or by group of several tensors (though this one is not very much tested beyond 2 tensors by split).
- Vibe coded.

* Create output directory if it doesn't exist in llama-quantize

* Create output directory if it doesn't exist in gguf-split

* Add exit when directory fails to be created on Windows

* Use std::filesystem

* cleanup

Display the size of the tensors overriden during the tensor loading (ikawrakow#1318)

* Display the size of the tensors overriden during the tensor loading

Ex:

`Tensor blk.60.ffn_gate_exps.weight buffer type overriden to CPU
Tensor blk.60.ffn_up_exps.weight buffer type overriden to CPU`

become

`Tensor blk.60.ffn_up_exps.weight (size = 668467200 bytes) buffer type overriden to CPU
Tensor blk.60.ffn_gate_exps.weight (size = 668467200 bytes) buffer type overriden to CPU`

And pass in debug the later displayed size of the unnamed buffer overrides.

Ex : `llm_load_tensors:        CPU buffer size =   XXX.XX MiB`

That double display is cluttering the screen without being very informative.

* change bytes display to MiB.

Co-authored-by: Kawrakow <iwankawrakow@gmail.com>

---------

Co-authored-by: Kawrakow <iwankawrakow@gmail.com>

Fused delta-net (ikawrakow#1315)

* Revive fused delta-net

* Add command line argument for fused delta net

* Simplify/improve CUDA delta-net

* Add -fdn to llama-bench

* More CUDA fused delta net optimizations

* CPU optimizations

* Much faster fused delta-net on the CPU

It seems it is faster than the chunked implementation!

* Change meaning of fdn from bool flag to threshold value

* Use eps = 1e-6

* Give some nodes a name

Fix KT quantization yet again (ikawrakow#1321)

* Fix KT quantization yet again

* Add same 1e-16f check for all quants in iqk_uantize.cpp

* Fixes for k-quants

* Also this one

server: enable checkpoint for recurrent models (ikawrakow#1310)

* server: enable checkpoint for recurrent models

create checkpoint after cancel

fix ban string and rm context during rewind

add checkpoint interval

only save recurrent cache

* save checkpoint during pp

---------

Co-authored-by: firecoperana <firecoperana>

Faster quantization for MoE models with many experts (ikawrakow#1322)

Fused delta net 2 (ikawrakow#1320)

* Revive fused delta-net

* Add command line argument for fused delta net

* Simplify/improve CUDA delta-net

* Add -fdn to llama-bench

* More CUDA fused delta net optimizations

* CPU optimizations

* Much faster fused delta-net on the CPU

It seems it is faster than the chunked implementation!

* Change meaning of fdn from bool flag to threshold value

* Use eps = 1e-6

* Give some nodes a name

* Don't re-apply L2 norm - it has already been done

* This seems quite a bit better

* More tweaks

* Restore per context buffer size log

Not everybody uses models split in 2000 parts, and those who do,
actually want to see the biffer sizes.

iAdding support for dense Qwen-3.5 models (ikawrakow#1326)

add directio to llama-bench
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants