Displaying 1 to 30 of 74 repositories
Multimodal LLM with video, audio, image, and text understanding for enterprise applications
1d
254
Multilingual MoE text embedding model with 768D vectors, 100 languages, 512 token context
2d
302
State-of-the-art multilingual MoE text embedding model for retrieval across ~100 languages
2d
99
Efficient 284B MoE language model with 1M token context and multi-mode reasoning capabilities
2d
1.1K
Multimodal AI model with 35B MoE architecture for coding agents, reasoning, and vision tasks
7d
10K+
1
1T MoE multimodal agentic model with long-horizon coding, swarm orchestration, and native vision
8d
6.3K
119B MoE model with switchable reasoning mode, multimodal vision, and 256k context window
9d
1.3K
119B parameter hybrid model with reasoning, vision, and code capabilities (1M token context)
10d
1.8K
Multimodal LLM with 35B parameters for coding, agentic tasks, and vision-language understanding
13d
3.7K
Gemma 4: multimodal open AI models by Google, optimized for reasoning, coding, and long context.
27d
10K+
397B MoE model with 17B activation for reasoning, coding, agents, and multimodal understanding
1m
100K+
7
397B-parameter MoE multimodal LLM with 17B active params, 262K context, 201 languages
1m
10K+
1
744B MoE language model with 40B active params for reasoning, coding, and agentic tasks (FP8)
2m
10K+
3
Advanced coding agent model with 80B params (3B active MoE) for code generation and debugging
3m
10K+
1
Efficient 80B MoE coding model with 3B activated params, 256K context, and agentic capabilities
3m
50K+
1
Image generation model, uses a base latent diffusion model plus a refiner.
3m
10K+
7
GLM-4.7-Flash is a top 30B-A3B MoE, balancing strong performance with efficient deployment.
3m
10K+
4
GLM-4.7-Flash is a top 30B-A3B MoE, balancing strong performance with efficient deployment.
3m
10K+
3
Devstral Small 2 is an FP8 instruct LLM for agentic SWE tasks, codebase tooling, and SWE-bench.
4m
10K+
4
FunctionGemma is a 270M open model for fine-tuned, offline function-calling agents on small devices.
4m
6.2K
1
FunctionGemma is a 270M open model for fine-tuned, offline function-calling agents on small devices.
4m
9.6K
2
Kimi K2 Thinking: open-source agent with deep reasoning, stable tool use, fast INT4, 256k context.
5m
50K+
2
Kimi K2 Thinking: open-source agent with deep reasoning, stable tool use, fast INT4, 256k context.
5m
10K+
2
DeepSeek-V3.2 boosts efficiency and reasoning with DSA, scalable RL, agentic data—IMO/IOI wins.
5m
50K+
10
Ministral 3: compact vision-enabled model with near-24B performance, optimized for local edge use
5m
10K+
4
Ministral 3: compact vision-enabled model with near-24B performance, optimized for local edge use
5m
50K+
2
Multilingual reranking model for text retrieval, scoring document relevance across 119 languages.
5m
10K+
3
Multilingual reranking model for text retrieval, scoring document relevance across 119 languages.
5m
10K+