We train and serve models that help write better code.

Morph does pre-training continuation, RL, and specialized inference engines. Every model we ship is trained, evaluated, and deployed in-house.

Many have said the future is small, specialized models. We only partially agree.

The future is small, specialized models with specialized inference, and we're building both vertically. They apply edits, search codebases, compress context, and review PRs. Each task gets its own model, its own training data, its own evaluation suite.

The result: sub-second edit application at 10,500 tok/s. Codebase search in under 6 seconds. Context compaction at 33,000 tok/sec. These aren't benchmarks on a leaderboard. They're production numbers from teams shipping code every day.

Work with us.

If you're a passionate ML engineer looking to work on a very small team and push to production every day, we want to hear from you.

View Open Roles

Y Combinator S23