Welcome to Local-Codex, the enterprise-ready solution for AI-powered development assistance! This guide will help you get up and running with minimal fuss.
For Jetson Users Only If you're working on NVIDIA Jetson hardware:
- Clone the jetson-containers repository (if not already installed):
git clone https://github.com/dusty-nv/jetson-containers
- This provides the necessary image for running on Jetson platforms
- Skip this step if not using Jetson hardware
Local-Codex works with any Chat Completions API provider:
- ✅ Ollama (default in our docker setup)
- ✅ vLLM
- ✅ OpenAI
- ✅ Groq
- ✅ Any other Chat Completions API compatible service Choose what works best for your home or organization's needs!
Our Docker Compose includes two containers:
- Ollama Container: Runs the LLM service
- Can be replaced with any other Chat Completions API
- Skip this container if using an external API
- Python Container: Core application
- Installs Local-Codex
- Installs ORS (OpenAI Responses Server)
- ORS adapts Chat Completions API to Responses API format
For the best experience, we recommend installing locally:
Once installed, you're all set to supercharge your development workflow!
Happy coding! 💪
TeaBranch. (2025). openai-responses-server: Open-source server the serves any AI provider with OpenAI ChatCompletions as OpenAI's Responses API and hosted tools. [Computer software]. GitHub. https://github.com/teabranch/openai-responses-server
cff-version: 1.2.0
title: >-
Jetson Containers(Machine Learning Containers for Jetson and JetPack)
message: >-
If you use this software, please cite it using the
metadata from this file.
type: software
authors:
- given-names: Dustin
family-names: Franklin
affiliation: Nvidia
repository-code: 'https://github.com/dusty-nv/jetson-containers'
url: 'https://www.jetson-ai-lab.com/'
abstract: Machine Learning Containers for Jetson and JetPack
license: MIT