On-device AI coding.
No cloud. No limits.

A complete coding agent that executes entirely on your machine. No API calls. No usage caps.

Zero telemetry Native inference 100% offline Your hardware, your rules No tokens, no limits Specialized SLM Unbounded context Zero telemetry Native inference 100% offline Your hardware, your rules No tokens, no limits Specialized SLM Unbounded context
The problem

You don't own your AI.
And you're being watched.

MONITORING ACTIVE TELEMETRY
Data extraction 001

They train on your code.

Every prompt. Every file. Every fix.
It flows through infrastructure you don't control — improving systems they want to use to replace you.

Artificial scarcity 002

They meter your ambition.

Slowdowns, overages, caps.
Right when you're deep in a sprint, the meter decides you've had enough.

Silent downgrades 003

They change the model.

They silently downgrade to cheaper models during peak load. Full price, degraded experience.

Cloud dependency 004

They control your flow.

Every completion makes a round trip across the internet.
Thousands of tiny interruptions, every single day.

Introducing Rig

Everything local.
Own your AI.

A complete AI coding agent running entirely on your own hardware. No usage limits. No cloud dependency.

YOUR MACHINE YOUR CODE KEYSTROKES · FILES RIG ✓ LOCAL INFERENCE GPU INDEX MODEL RESPONSE <300ms · ON DEVICE CLOUD TELEMETRY
Cloud servers
Severed
Your machine
Rig model active
Severed
Nothing leaves
CLOUD API TELEMETRY BLOCKED BLOCKED FIREWALL YOUR MACHINE RIG ✓ MODEL ACTIVE GPU INDEX NOTHING LEAVES DATA STAYS LOCAL
Offline

Work offline

Flights. Spotty Wi-Fi. Network outages. Nothing stops your flow.

Unlimited

Remove the meter

Refactor the whole codebase. Riff on an idea all day. Run agent loops without thinking about cost.

Privacy

Sever the connection

Your code, keystrokes, and files never leave your machine. Not anonymized. Not aggregated. Not sent.

Latency

Stop waiting

No round-trip to a data center. Inference happens on your machine, in single-digit milliseconds.

Our Approach

Purpose beats scale.

Rig is a closed system — model, context, tools, and inference — engineered together for one job: real coding work.

Training Focus
Parameters dedicated to code

Rig            ████████████████████ 100%
Most AI models ████░░░░░░░░░░░░░░░░ ~15–20%

General-purpose models spread capacity
across chat, translation, creative
writing, and more.

Rig dedicates every parameter
to engineering.
Model size (memory required)

Cloud models  ████████████████████ 200+ GB
Open source   ██████░░░░░░░░░░░░░░ 28–140 GB
Rig           ░░░░░░░░░░░░░░░░░░░ 8 GB

Fits in 16 GB unified memory.
Accuracy loss: <0.3%
First token latency

Rig           ░░░░░░░░░░░░░░░░░░░ 300 ms
Cloud APIs    ████████████████████ 400–1,000 ms

Cost per 1K tokens

Rig                               $0.00
Cloud APIs    ████████████████████ $0.01–0.06
Capabilities

Your machine, unleashed.

[ 01 ]

Understands your architecture.

Builds a connected model of modules, dependencies, and relationships so reasoning happens across files and aligns with your architecture.

[ 02 ]

Tracks relationships, prevents breakage.

Edits that respect function contracts, type boundaries, and dependency graphs — reducing bugs and regressions.

[ 03 ]

Strategizes before acting.

Explore → Plan → Execute workflows ensure multiple steps are reasoned out before changes occur.

[ 04 ]

Executes complex coding workflows.

From refactors to test generation to feature builds — coordinate tools, code edits, web search, and commands as needed.

[ 05 ]

Isolates agent sandboxes.

Each agent runs in its own workspace so experiments are safe, parallel workflows don't clash, and code changes stay isolated until you merge them.

[ 06 ]

Runs at full speed.

Custom Rust inference engine optimized for CUDA and Metal — delivering up to 144 tokens per second on consumer hardware.

Latency 0ms No round-trip required
Privacy 100% Air-gapped by design
Cost / Token $0 Your GPU, your tokens
Uptime Local No dependency on cloud
Engineered intelligence

Built for control freaks

Custom Model
Optimized for consumer hardware
Inference
Cross-OS using Rust
Context Graph
Repo-wide code understanding
Terminal UI
Built in Rust and blazing fast
Heavily Tuned
Consistent tool calls and plan use
Opinionated
Focused on code correctness
rig://localhost · offline
λ rig init
  ██████╗  ██╗  ██████╗
  ██╔══██╗ ██║ ██╔════╝
  ██████╔╝ ██║ ██║ ███╗
  ██╔══██╗ ██║ ██║  ██║
  ██║  ██║ ██║ ╚██████║
  ╚═╝  ╚═╝ ╚═╝  ╚═════╝
> Scanning hardware...
> Found M4 · 16GB RAM
> Loading RIG Model OK
> Indexing 2,418 files · 87,102 symbols
Ready. Network: OFF · Telemetry: OFF
λ explain this regex to
Neural Engine
RG-800 Local Ops
Custom Model
Optimized for consumer hardware
Terminal UI
Built in Rust and blazing fast
Inference
Cross-OS using Rust
Heavily Tuned
Consistent tool calls and plan use
Context Graph
Repo-wide code understanding
Opinionated
Focused on code correctness
Early access

Rig is almost ready.

We're inviting engineers to run it on real code and help shape what ships.

FAQ

Frequently asked questions.

01What is Rig?

Rig is a local-first AI coding assistant that runs entirely on your machine. It uses a modified open-source model post-trained exclusively for code, executed by a custom Rust inference engine optimized for Apple Silicon. Rig delivers up to 144 tokens per second with 300ms first-token latency, requires no API calls, has no usage caps, collects zero telemetry, and costs $0 per token. All code and files stay on your machine. Rig currently supports macOS with Linux and Windows support planned.

02What Model does Rig use?

Rig is a custom trained model based on an open source model. We extensively trained it to work exclusively with the Rig agent harness, context engine, and tools. This allows us to shrink the model's total footprint without losing intelligence or coding capability.

03What are the Hardware Requirements?

Rig is optimized to run on Apple Silicon devices using M2 or later with at least 16GB of RAM. Support for Window and Linux are coming soon.

04How Does Rig Compare to Large Cloud Models?

Rig's model is still in development so we do not have benchmarks available yet. Our early tests indicate the Rig model will be on par with state of the art models thanks to the combination of our context engine and post training pipeline.

05Can Rig Search the Web?

Yes, Rig has all the same tools you'd expect from a coding agent, including web search, file read / write, plan mode, and more.

06How will Rig be Priced?

Rig's pricing model is planned to be a flat monthly or annual subscription on par with other coding agents but completely unlimited and offline.

07Will Rig Collect my Data?

No, Rig is committed to being the most secure and private coding agent available. Our telemetry will be limited to a license check with a grace period. Your code and conversations will never leave your machine.

08When Will Rig be Available?

Soon! We are planning to start rolling out to our waitlist in Q2 2026 with a broader release targeted for later this year. We're focused on creating the best possible coding assistant capable of supporting real software engineers on their most important projects.

Break free from big AI

No credit card. No usage meter.