Track usage and reset windows across Anthropic, Codex, Synthetic, Z.ai, Copilot, MiniMax, and Antigravity. Route work before limits hit. Detect anomalies. Monitor burn rates. All data stays on your machine.
$ curl -fsSL https://raw.githubusercontent.com/onllm-dev/onwatch/main/install.sh | bash
Copy
$ brew install onllm-dev/tap/onwatch
Copy
PS> irm https://raw.githubusercontent.com/onllm-dev/onwatch/main/install.ps1 | iex
Copy
Live data from public GitHub stargazer profiles. Country spread, organizations, and preference at a glance.
Interactive map is unavailable right now. Country coverage is listed below.
Public profile metadata only. Organization names are examples and do not imply endorsement.
Snapshot is stale. Showing neutral trust mode until refreshed data is verified.
Curated trust snapshot unavailable. Showing neutral trust mode.
Different quotas, different reset cycles, different gotchas. onWatch normalizes all of it.
They show current usage. onWatch shows the full picture.
Charts your consumption across hours, days, and weeks. See when you burn quota fastest. Plan around it.
OpenAI has reset limits early multiple times. onWatch catches unexpected resets and rate limit changes before they block your work.
See remaining headroom across all providers. Anthropic at 90%? Codex still has capacity. Switch before you hit a wall.
Track burn rates and cycle utilization over time. When renewal comes, you know whether to upgrade, downgrade, or stay put. All data stays local.
Context, history, and projections that provider dashboards do not have.
Save profiles with onwatch codex profile save, switch in the dashboard. Per-account charts, sessions, and cycle history. Good for personal/work separation.
Time-series charts (1h to 30d) for every provider. Automatic reset cycle detection for Anthropic 5h/7d, Synthetic subscription, Z.ai daily budget. Peak usage and delta per cycle.
Live countdown to each quota reset. Extrapolates your burn rate to the next boundary - tells you if you will run out before relief arrives.
Every coding session logs peak consumption per provider. Compare sessions side-by-side and see utilization trends across cycles.
SMTP email or browser push when quotas cross your thresholds. AES-256 encrypted credentials. Configure per-quota levels and delivery channels.
Zero telemetry - your usage data never leaves your machine. AES-256 encrypted credentials, constant-time auth, parameterized SQL.
Background daemon polls provider APIs every 60s. Stores snapshots in SQLite. Serves a dashboard. That is it.
macOS, Linux, and Windows. Pick what works for you.
Install via Homebrew or the one-line shell installer. Both auto-detect Claude Code and Codex credentials. Homebrew handles PATH and updates natively; the shell installer sets up systemd (Linux) or self-daemonizes (macOS).
PowerShell installer for Windows with interactive setup and auto-detection of Claude Code/Codex credentials. Docker for containerized deployments with distroless image (~12 MB), non-root user, and persistent data volume.
Download the binary for your platform from GitHub Releases. Best for Linux with systemd service management.
Clone the repo and build with app.sh. Requires Go 1.25+. Full control over build flags and development workflow. Includes 486 tests with race detection.
See DEVELOPMENT.md for advanced build options and cross-compilation.
The fastest way is Homebrew: brew install onllm-dev/tap/onwatch, then onwatch setup to configure your providers interactively. Alternatively, install with one command: curl -fsSL https://raw.githubusercontent.com/onllm-dev/onwatch/main/install.sh | bash. The setup wizard auto-detects Claude Code and Codex credentials, prompts for API keys, and configures dashboard credentials. onWatch polls each configured provider every 60 seconds, stores snapshots in SQLite, and serves a dashboard at localhost:9211 with live countdowns, charts, and cross-provider views.
Yes. onWatch monitors the API provider (Synthetic, Z.ai, Anthropic, Codex, GitHub Copilot, MiniMax, or Antigravity), not the coding tool. MiniMax support is live for MiniMax Coding Plan accounts on platform.minimax.io. Any tool that uses these API keys-including Cline, Roo Code, Kilo Code, Claude Code, Cursor, GitHub Copilot, MiniMax, Antigravity, and others-will have its usage tracked automatically.
Anthropic's Pro/Max plan exposes utilization percentages and reset times for five_hour and seven_day windows, plus per-model breakdowns (seven_day_sonnet, seven_day_opus). onWatch polls this data, stores historical snapshots, and adds what Anthropic doesn't show: usage trends over time, reset cycle detection, rate projections, and cross-provider context alongside Synthetic, Z.ai, Codex, GitHub Copilot, MiniMax, and Antigravity. Set ANTHROPIC_TOKEN in your .env or let onWatch auto-detect from Claude Code credentials.
Antigravity provides access to multiple AI models (Claude, Gemini, GPT). onWatch auto-detects the Antigravity language server running on your machine by scanning for the process and extracting connection details. Set ANTIGRAVITY_ENABLED=true in your .env file. Models are grouped into logical quota pools (Claude+GPT, Gemini Pro, Gemini Flash) for cleaner tracking. For Docker deployments, configure ANTIGRAVITY_BASE_URL and ANTIGRAVITY_CSRF_TOKEN manually.
No. onWatch has zero telemetry. All usage data is stored locally in a SQLite file on your machine. The only outbound network calls are to the Synthetic, Z.ai, Anthropic, Codex, GitHub Copilot, MiniMax, and Antigravity quota APIs you configure. No analytics, no tracking, no cloud. The source code is fully auditable on GitHub (GPL-3.0).
onWatch uses <60 MB RAM under all conditions (typically ~34 MB idle, ~43 MB under heavy load), measured with all seven providers (Synthetic, Z.ai, Anthropic, Codex, GitHub Copilot, MiniMax, Antigravity) polling in parallel. Breakdown: Go runtime (5 MB), SQLite in-process (2 MB), HTTP server (1 MB), polling buffer (1 MB). This is lighter than a single browser tab and designed to run as a background daemon indefinitely.
One command to install. All data stays local. Free, open source, zero telemetry.