LLM Gateway
Guides

Codex CLI Integration

Use any model with OpenAI's Codex CLI through LLM Gateway. One config file, full cost tracking.

Codex CLI is OpenAI's open-source terminal coding agent. By default it connects to OpenAI's API, but with LLM Gateway you can route it through a single gateway—use GPT-5.3 Codex, Gemini, Claude, or any of 180+ models while keeping full cost visibility.

One config file. No code changes. Full cost tracking in your dashboard.

Setup

Sign Up for LLM Gateway

Sign up free — no credit card required. Copy your API key from the dashboard.

Set Your API Key

Set your LLM Gateway API key as the OpenAI key:

export OPENAI_API_KEY=llmgtwy_your_api_key_here

Create Config File

Create or edit ~/.codex/config.toml:

openai_base_url = "https://api.llmgateway.io/v1"
model = "auto"
model_reasoning_effort = "high"

[tui]
show_tooltips = false

[model_providers.openai]
name = "OpenAI"
base_url = "https://api.llmgateway.io/v1"

Run Codex CLI

codex

All requests will now be routed through LLM Gateway.

Why This Works

LLM Gateway's /v1 endpoint is fully OpenAI-compatible. Codex CLI sends requests to our gateway instead of OpenAI directly, and we route them to the right provider behind the scenes. This means:

  • Use any model — GPT-5.3 Codex, Gemini, Claude, or 180+ others
  • Keep your workflow — Codex CLI doesn't know the difference
  • Track costs — Every request appears in your LLM Gateway dashboard
  • Automatic caching — Repeated requests hit cache, saving money

Configuration Explained

Base URL

The openai_base_url and base_url fields point Codex CLI to LLM Gateway instead of OpenAI:

openai_base_url = "https://api.llmgateway.io/v1"

Model Selection

Use auto to let LLM Gateway pick the best model, or set a specific one from the models page:

model = "auto"
# or pick a specific model
model = "gpt-5.3-codex"

Reasoning Effort

Control how much reasoning the model uses. Options are low, medium, and high:

model_reasoning_effort = "high"

Choosing Models

Use auto to let LLM Gateway pick the best model automatically, or choose a specific one from the models page:

# let LLM Gateway pick the best model
model = "auto"

# or pick a specific model
model = "gpt-5.3-codex"

What You Get

  • Any model in Codex CLI — GPT-5.3 Codex for heavy lifting, lighter models for routine tasks
  • Cost visibility — See exactly what each coding session costs
  • One bill — Stop managing separate accounts for OpenAI, Anthropic, Google
  • Response caching — Repeated requests hit cache automatically
  • Discounts — Check discounted models for savings up to 90%

Troubleshooting

Authentication errors

Make sure your OPENAI_API_KEY environment variable is set to your LLM Gateway API key (starts with llmgtwy_).

Model not found

Verify the model ID matches exactly what's listed on the models page. Model IDs are case-sensitive.

Connection issues

Check that base_url is set to https://api.llmgateway.io/v1 (note the /v1 at the end).

View all available models on the models page.

Need help? Join our Discord community for support and troubleshooting assistance.

How is this guide?

Last updated on