zai-org/GLM-5
TEXT GENERATIONConcurrency Cost:4Model Size:754BQuant:FP8Ctx Length:32kPublished:Feb 11, 2026License:mitArchitecture:Transformer1.9K Open Weights Warm
GLM-5 is a large language model developed by zai-org, scaling from 355B (32B active) to 744B (40B active) parameters and trained on 28.5T tokens. It integrates DeepSeek Sparse Attention (DSA) for efficient long-context capacity and utilizes a novel asynchronous RL infrastructure called 'slime' for improved post-training. GLM-5 is designed for complex systems engineering and long-horizon agentic tasks, achieving best-in-class performance among open-source models in reasoning, coding, and agentic benchmarks.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–