Codex Harness
The bundledcodex plugin lets OpenClaw run embedded agent turns through the
Codex app-server instead of the built-in PI harness.
Use this when you want Codex to own the low-level agent session: model
discovery, native thread resume, native compaction, and app-server execution.
OpenClaw still owns chat channels, session files, model selection, tools,
approvals, media delivery, and the visible transcript mirror.
The harness is off by default. It is selected only when the codex plugin is
enabled and the resolved model is a codex/* model, or when you explicitly
force embeddedHarness.runtime: "codex" or OPENCLAW_AGENT_RUNTIME=codex.
If you never configure codex/*, existing PI, OpenAI, Anthropic, Gemini, local,
and custom-provider runs keep their current behavior.
Pick the right model prefix
OpenClaw has separate routes for OpenAI and Codex-shaped access:| Model ref | Runtime path | Use when |
|---|---|---|
openai/gpt-5.4 | OpenAI provider through OpenClaw/PI plumbing | You want direct OpenAI Platform API access with OPENAI_API_KEY. |
openai-codex/gpt-5.4 | OpenAI Codex OAuth provider through PI | You want ChatGPT/Codex OAuth without the Codex app-server harness. |
codex/gpt-5.4 | Bundled Codex provider plus Codex harness | You want native Codex app-server execution for the embedded agent turn. |
codex/* model refs. Existing openai/*,
openai-codex/*, Anthropic, Gemini, xAI, local, and custom provider refs keep
their normal paths.
Requirements
- OpenClaw with the bundled
codexplugin available. - Codex app-server
0.118.0or newer. - Codex auth available to the app-server process.
OPENAI_API_KEY, plus
optional Codex CLI files such as ~/.codex/auth.json and
~/.codex/config.toml. Use the same auth material your local Codex app-server
uses.
Minimal config
Usecodex/gpt-5.4, enable the bundled plugin, and force the codex harness:
plugins.allow, include codex there too:
agents.defaults.model or an agent model to codex/<model> also
auto-enables the bundled codex plugin. The explicit plugin entry is still
useful in shared configs because it makes the deployment intent obvious.
Add Codex without replacing other models
Keepruntime: "auto" when you want Codex for codex/* models and PI for
everything else:
/model codexor/model codex/gpt-5.4uses the Codex app-server harness./model gptor/model openai/gpt-5.4uses the OpenAI provider path./model opususes the Anthropic provider path.- If a non-Codex model is selected, PI remains the compatibility harness.
Codex-only deployments
Disable PI fallback when you need to prove that every embedded agent turn uses the Codex harness:codex/* ref, the app-server is too old, or the
app-server cannot start.
Per-agent Codex
You can make one agent Codex-only while the default agent keeps normal auto-selection:/new creates a fresh
OpenClaw session and the Codex harness creates or resumes its sidecar app-server
thread as needed. /reset clears the OpenClaw session binding for that thread.
Model discovery
By default, the Codex plugin asks the app-server for available models. If discovery fails or times out, it uses the bundled fallback catalog:codex/gpt-5.4codex/gpt-5.4-minicodex/gpt-5.2
plugins.entries.codex.config.discovery:
App-server connection and policy
By default, the plugin starts Codex locally with:appServer fields:
| Field | Default | Meaning |
|---|---|---|
transport | "stdio" | "stdio" spawns Codex; "websocket" connects to url. |
command | "codex" | Executable for stdio transport. |
args | ["app-server", "--listen", "stdio://"] | Arguments for stdio transport. |
url | unset | WebSocket app-server URL. |
authToken | unset | Bearer token for WebSocket transport. |
headers | {} | Extra WebSocket headers. |
requestTimeoutMs | 60000 | Timeout for app-server control-plane calls. |
approvalPolicy | "never" | Native Codex approval policy sent to thread start/resume/turn. |
sandbox | "workspace-write" | Native Codex sandbox mode sent to thread start/resume. |
approvalsReviewer | "user" | Use "guardian_subagent" to let Codex guardian review native approvals. |
serviceTier | unset | Optional Codex service tier, for example "priority". |
OPENCLAW_CODEX_APP_SERVER_BINOPENCLAW_CODEX_APP_SERVER_ARGSOPENCLAW_CODEX_APP_SERVER_APPROVAL_POLICYOPENCLAW_CODEX_APP_SERVER_SANDBOXOPENCLAW_CODEX_APP_SERVER_GUARDIAN=1
Common recipes
Local Codex with default stdio transport:codex/* model, provider, approval policy, sandbox, and service tier to
app-server again. Switching from codex/gpt-5.4 to codex/gpt-5.2 keeps the
thread binding but asks Codex to continue with the newly selected model.
Codex command
The bundled plugin registers/codex as an authorized slash command. It is
generic and works on any channel that supports OpenClaw text commands.
Common forms:
/codex statusshows live app-server connectivity, models, account, rate limits, MCP servers, and skills./codex modelslists live Codex app-server models./codex threads [filter]lists recent Codex threads./codex resume <thread-id>attaches the current OpenClaw session to an existing Codex thread./codex compactasks Codex app-server to compact the attached thread./codex reviewstarts Codex native review for the attached thread./codex accountshows account and rate-limit status./codex mcplists Codex app-server MCP server status./codex skillslists Codex app-server skills.
/codex resume writes the same sidecar binding file that the harness uses for
normal turns. On the next message, OpenClaw resumes that Codex thread, passes the
currently selected OpenClaw codex/* model into app-server, and keeps extended
history enabled.
The command surface requires Codex app-server 0.118.0 or newer. Individual
control methods are reported as unsupported by this Codex app-server if a
future or custom app-server does not expose that JSON-RPC method.
Tools, media, and compaction
The Codex harness changes the low-level embedded agent executor only. OpenClaw still builds the tool list and receives dynamic tool results from the harness. Text, images, video, music, TTS, approvals, and messaging-tool output continue through the normal OpenClaw delivery path. When the selected model uses the Codex harness, native thread compaction is delegated to Codex app-server. OpenClaw keeps a transcript mirror for channel history, search,/new, /reset, and future model or harness switching. The
mirror includes the user prompt, final assistant text, and lightweight Codex
reasoning or plan records when the app-server emits them.
Media generation does not require PI. Image, video, music, PDF, TTS, and media
understanding continue to use the matching provider/model settings such as
agents.defaults.imageGenerationModel, videoGenerationModel, pdfModel, and
messages.tts.
Troubleshooting
Codex does not appear in/model: enable plugins.entries.codex.enabled,
set a codex/* model ref, or check whether plugins.allow excludes codex.
OpenClaw falls back to PI: set embeddedHarness.fallback: "none" or
OPENCLAW_AGENT_HARNESS_FALLBACK=none while testing.
The app-server is rejected: upgrade Codex so the app-server handshake
reports version 0.118.0 or newer.
Model discovery is slow: lower plugins.entries.codex.config.discovery.timeoutMs
or disable discovery.
WebSocket transport fails immediately: check appServer.url, authToken,
and that the remote app-server speaks the same Codex app-server protocol version.
A non-Codex model uses PI: that is expected. The Codex harness only claims
codex/* model refs.