FLUX.2 Image Generation in under 1 second. Read More →

Model Library

Browse our library of open source models that are ready to deploy behind an API endpoint in seconds.

All
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
🔥 Trending models
DeepSeek V3.2
Live
DeepSeek V3.2
LLM
FLUX.2 Dev
Live
FLUX.2 Dev
Image
FLUX.2 [klein] 9B
Live
FLUX.2 [klein] 9B
Image
GLM-5
GLM-5
LLM
Kimi K2.5
Live
Kimi K2.5
LLM
Vision
MiniMax M2.5
MiniMax M2.5
LLM
DeepSeek R1
DeepSeek R1
LLM
DeepSeek R1-0528
DeepSeek R1-0528
LLM
DeepSeek R1 Distill Llama 70B
DeepSeek R1 Distill Llama 70B
LLM
DeepSeek V3
Live
DeepSeek V3
LLM
DeepSeek V3.1
Live
DeepSeek V3.1
LLM
DeepSeek V3.2
Live
DeepSeek V3.2
LLM
Audio
GLM-4.7
GLM-4.7
LLM
Vision
Audio
Qwen3-Omni-30B-A3B
Qwen3-Omni-30B-A3B
LLM
Audio
Vision
Vision
Gemma 3 12B
Gemma 3 12B
LLM
Vision
Gemma 3 27B
Gemma 3 27B
LLM
Vision
GLM-4.7
GLM-4.7
LLM
Vision
Audio
Kimi K2.5
Live
Kimi K2.5
LLM
Vision
Llama 4 Maverick
Llama 4 Maverick
LLM
Vision
Llama 4 Scout
Llama 4 Scout
LLM
Vision
Image
FLUX.2 Dev
Live
FLUX.2 Dev
Image
FLUX.2 [klein] 9B
Live
FLUX.2 [klein] 9B
Image
FLUX.2 [klein] 9B Fast
Live
FLUX.2 [klein] 9B Fast
Image
Image
DeepSeek R1
DeepSeek R1
LLM
DeepSeek R1-0528
DeepSeek R1-0528
LLM
DeepSeek R1 Distill Llama 70B
DeepSeek R1 Distill Llama 70B
LLM
DeepSeek V3
Live
DeepSeek V3
LLM
DeepSeek V3.1
Live
DeepSeek V3.1
LLM
DeepSeek V3.2
Live
DeepSeek V3.2
LLM
EXAONE 4.0 32B
EXAONE 4.0 32B
LLM
FLUX.2 Dev
Live
FLUX.2 Dev
Image
FLUX.2 [klein] 9B
Live
FLUX.2 [klein] 9B
Image
FLUX.2 [klein] 9B Fast
Live
FLUX.2 [klein] 9B Fast
Image
Gemma 3 12B
Gemma 3 12B
LLM
Vision
Gemma 3 27B
Gemma 3 27B
LLM
Vision
GLM-4.7
GLM-4.7
LLM
Vision
Audio
GLM-5
GLM-5
LLM
gpt-oss-120b
gpt-oss-120b
LLM
gpt-oss-20b
gpt-oss-20b
LLM
Kimi K2.5
Live
Kimi K2.5
LLM
Vision
Llama 3.1 405B
Llama 3.1 405B
LLM
Llama 3.3 70B
Llama 3.3 70B
LLM
Llama 4 Maverick
Llama 4 Maverick
LLM
Vision
Llama 4 Scout
Llama 4 Scout
LLM
Vision
MiMo-V2-Flash
MiMo-V2-Flash
LLM
MiniMax M2.5
MiniMax M2.5
LLM
Ministral 14B
Ministral 14B
LLM
Vision
Mistral Large 3
Mistral Large 3
LLM
Vision
Mistral Small 3.1 24B
Mistral Small 3.1 24B
LLM
Vision
Nemotron 3 Nano
Nemotron 3 Nano
LLM
Phi-4
Phi-4
LLM
Qwen2.5 72B
Qwen2.5 72B
LLM
Qwen3-235B-A22B
Qwen3-235B-A22B
LLM
Qwen3 30B-A3B
Qwen3 30B-A3B
LLM
Qwen3.5-397B-A17B
Qwen3.5-397B-A17B
LLM
Vision
Qwen3-Coder-480B-A35B
Qwen3-Coder-480B-A35B
LLM
Qwen3-Omni-30B-A3B
Qwen3-Omni-30B-A3B
LLM
Audio
Vision
Qwen3-VL-30B-A3B
Qwen3-VL-30B-A3B
LLM
Vision
Qwen3-VL-4B
Qwen3-VL-4B
LLM
Vision
Qwen3-VL-8B
Qwen3-VL-8B
LLM
Vision

Get started with Modular

  • Request a demo

    Schedule a demo of Modular and explore a custom end-to-end deployment built around your models, hardware, and performance goals.

    • Distributed, large-scale online inference endpoints

    • Highest-performance to maximize ROI and latency

    • Deploy in Modular cloud or your cloud

    • View all features with a custom demo

    Book a demo

    Talk with our sales lead Jay!

    30min demo.  Evaluate with your workloads.  Ask us anything.

  • Talk to us!

    Book a demo for a personalized walkthrough of Modular in your environment. Learn how teams use it to simplify systems and tune performance at scale.

    • Custom 30 min walkthrough of our platform

    • Cover specific model or deployment needs

    • Flexible pricing to fit your specific needs

    Book a demo

    Talk with our sales lead Jay!

  • Start using MAX

    ( FREE )

    Run any open source model in 5 minutes, then benchmark it. Scale it to millions yourself (for free!).

  • Start using Mojo

    ( FREE )

    Install Mojo and get up and running in minutes. A simple install, familiar tooling, and clear docs make it easy to start writing code immediately.