Know your AI quota.
Before it runs out.

Track usage and reset windows across Anthropic, Codex, Synthetic, Z.ai, Copilot, MiniMax, and Antigravity. Route work before limits hit. Detect anomalies. Monitor burn rates. All data stays on your machine.

One-command install
Mac & Linux $ curl -fsSL https://raw.githubusercontent.com/onllm-dev/onwatch/main/install.sh | bash Copy
Homebrew $ brew install onllm-dev/tap/onwatch Copy
Windows PS> irm https://raw.githubusercontent.com/onllm-dev/onwatch/main/install.ps1 | iex Copy

v2.11.5 · Go · GPL-3.0 · Zero Telemetry · <60 MB RAM · <20 MB Binary

localhost:9211
Dashboard
Anthropic Synthetic Z.ai Antigravity All
Anthropic
Five-Hour
0%
2h 18m
Warning
Seven-Day
0%
4d 12h
Healthy
Sonnet (7-Day)
0%
4d 12h
Healthy
Usage over time - Last 6 hours
Five-Hour Seven-Day Sonnet

Who likes onWatch

Live data from public GitHub stargazer profiles. Country spread, organizations, and preference at a glance.

GitHub Stars (Live) -- Snapshot context as of --
Countries (Aggregate) -- Includes full spread + Unknown
Organizations (Aggregate) -- People with public organization metadata

Country spread

Unknown locations: --

Star growth

Loading country aggregates...

Starred by employees of

Loading organization aggregates...

Public profile metadata only. Organization names are examples and do not imply endorsement.

Seven providers. One dashboard.

Different quotas, different reset cycles, different gotchas. onWatch normalizes all of it.

Codex
5-hour rolling + weekly limits + review requests + multi-account support
Anthropic
5-hour + 7-day + per-model (Sonnet, Opus) with auto-detected credentials
Synthetic
Subscription (~5h cycle) + hourly search (250/hr) + tool call discounts (~24h)
Z.ai
Daily token budget + time limits + tool calls with normalized field names
GitHub Copilot
Premium interactions (monthly 300-1500) + chat + completions tracking
MiniMax
Daily coding plan quota + shared pool (M2, M2.1, M2.5) + burn rate forecast
Antigravity
Claude + GPT + Gemini Pro + Gemini Flash - zero-config auto-detection
Cross-Provider View
See all providers side-by-side. Compare headroom at a glance and route work to whichever provider has capacity.
Only in onWatch

What provider dashboards miss

They show current usage. onWatch shows the full picture.

Historical trends

Charts your consumption across hours, days, and weeks. See when you burn quota fastest. Plan around it.

Anomaly detection

OpenAI has reset limits early multiple times. onWatch catches unexpected resets and rate limit changes before they block your work.

Cross-provider routing

See remaining headroom across all providers. Anthropic at 90%? Codex still has capacity. Switch before you hit a wall.

Subscription intelligence

Track burn rates and cycle utilization over time. When renewal comes, you know whether to upgrade, downgrade, or stay put. All data stays local.

What you get

Context, history, and projections that provider dashboards do not have.

Multi-Account Codex Beta

Save profiles with onwatch codex profile save, switch in the dashboard. Per-account charts, sessions, and cycle history. Good for personal/work separation.

Usage trends & cycle detection

Time-series charts (1h to 30d) for every provider. Automatic reset cycle detection for Anthropic 5h/7d, Synthetic subscription, Z.ai daily budget. Peak usage and delta per cycle.

Live countdowns & projections

Live countdown to each quota reset. Extrapolates your burn rate to the next boundary - tells you if you will run out before relief arrives.

Session tracking & insights

Every coding session logs peak consumption per provider. Compare sessions side-by-side and see utilization trends across cycles.

Email & push alerts Beta

SMTP email or browser push when quotas cross your thresholds. AES-256 encrypted credentials. Configure per-quota levels and delivery channels.

Enterprise-grade security

Zero telemetry - your usage data never leaves your machine. AES-256 encrypted credentials, constant-time auth, parameterized SQL.

How it works

Background daemon polls provider APIs every 60s. Stores snapshots in SQLite. Serves a dashboard. That is it.

Your Providers
Synthetic + Z.ai + Anthropic + Codex + Copilot
onWatch Agent
Poll → Detect → Store → Analyze
Intelligence Dashboard
Patterns, projections, decisions
<60 MB
RAM Usage
0
Dependencies
SQLite
Local Intelligence Store
60s
Data Collection Interval

Install

macOS, Linux, and Windows. Pick what works for you.

Windows & Docker

PowerShell installer for Windows with interactive setup and auto-detection of Claude Code/Codex credentials. Docker for containerized deployments with distroless image (~12 MB), non-root user, and persistent data volume.

Windows (PowerShell)
# Interactive installer - auto-detects credentials irm https://raw.githubusercontent.com/onllm-dev/onwatch/main/install.ps1 | iex # Manage: onwatch, onwatch stop, onwatch --debug
Docker
# Clone & configure git clone https://github.com/onllm-dev/onwatch.git cd onwatch && cp .env.docker.example .env vim .env # add your API keys # Run with Docker Compose docker-compose up -d docker-compose logs -f # view logs
Manual Download

Download the binary for your platform from GitHub Releases. Best for Linux with systemd service management.

download & install
# Download (Linux AMD64) curl -L -o onwatch \ https://github.com/onllm-dev/onwatch/releases/latest/download/onwatch-linux-amd64 chmod +x onwatch && sudo mv onwatch /usr/local/bin/
configure & manage (Linux)
onwatch setup # configure providers # Manage via systemd systemctl start onwatch # start systemctl stop onwatch # stop systemctl status onwatch # check status journalctl -u onwatch -f # live logs
Build from Source

Clone the repo and build with app.sh. Requires Go 1.25+. Full control over build flags and development workflow. Includes 486 tests with race detection.

clone & build
git clone https://github.com/onllm-dev/onwatch.git cd onwatch cp .env.example .env vim .env # add your API keys ./app.sh --build # build binary
develop & test
./app.sh --test # run tests with -race ./app.sh --smoke # quick pre-commit check ./onwatch --debug # run in foreground ./onwatch # start daemon

See DEVELOPMENT.md for advanced build options and cross-compilation.

Frequently asked questions

How do I get started?

The fastest way is Homebrew: brew install onllm-dev/tap/onwatch, then onwatch setup to configure your providers interactively. Alternatively, install with one command: curl -fsSL https://raw.githubusercontent.com/onllm-dev/onwatch/main/install.sh | bash. The setup wizard auto-detects Claude Code and Codex credentials, prompts for API keys, and configures dashboard credentials. onWatch polls each configured provider every 60 seconds, stores snapshots in SQLite, and serves a dashboard at localhost:9211 with live countdowns, charts, and cross-provider views.

Does onWatch work with Cline, Roo Code, Kilo Code, or Claude Code?

Yes. onWatch monitors the API provider (Synthetic, Z.ai, Anthropic, Codex, GitHub Copilot, MiniMax, or Antigravity), not the coding tool. MiniMax support is live for MiniMax Coding Plan accounts on platform.minimax.io. Any tool that uses these API keys-including Cline, Roo Code, Kilo Code, Claude Code, Cursor, GitHub Copilot, MiniMax, Antigravity, and others-will have its usage tracked automatically.

How does Anthropic API tracking work?

Anthropic's Pro/Max plan exposes utilization percentages and reset times for five_hour and seven_day windows, plus per-model breakdowns (seven_day_sonnet, seven_day_opus). onWatch polls this data, stores historical snapshots, and adds what Anthropic doesn't show: usage trends over time, reset cycle detection, rate projections, and cross-provider context alongside Synthetic, Z.ai, Codex, GitHub Copilot, MiniMax, and Antigravity. Set ANTHROPIC_TOKEN in your .env or let onWatch auto-detect from Claude Code credentials.

How does Antigravity tracking work?

Antigravity provides access to multiple AI models (Claude, Gemini, GPT). onWatch auto-detects the Antigravity language server running on your machine by scanning for the process and extracting connection details. Set ANTIGRAVITY_ENABLED=true in your .env file. Models are grouped into logical quota pools (Claude+GPT, Gemini Pro, Gemini Flash) for cleaner tracking. For Docker deployments, configure ANTIGRAVITY_BASE_URL and ANTIGRAVITY_CSRF_TOKEN manually.

Does onWatch send any data to external servers?

No. onWatch has zero telemetry. All usage data is stored locally in a SQLite file on your machine. The only outbound network calls are to the Synthetic, Z.ai, Anthropic, Codex, GitHub Copilot, MiniMax, and Antigravity quota APIs you configure. No analytics, no tracking, no cloud. The source code is fully auditable on GitHub (GPL-3.0).

How much memory does onWatch use?

onWatch uses <60 MB RAM under all conditions (typically ~34 MB idle, ~43 MB under heavy load), measured with all seven providers (Synthetic, Z.ai, Anthropic, Codex, GitHub Copilot, MiniMax, Antigravity) polling in parallel. Breakdown: Go runtime (5 MB), SQLite in-process (2 MB), HTTP server (1 MB), polling buffer (1 MB). This is lighter than a single browser tab and designed to run as a background daemon indefinitely.

Stop flying blind on quota

One command to install. All data stays local. Free, open source, zero telemetry.