TL;DR

On 2026-04-23, OpenAI shipped Auto-review as a stable Codex feature in codex-cli 0.124.0. Eligible approval prompts are now routed through an automatic reviewer agent before the action runs. The Codex app shows a risk level and one of four statuses — Approved, Denied, Stopped, or Timed out — so developers can judge the AI's assessment before they click allow. The goal: longer, safer autonomous workflows with dramatically fewer manual approvals.

What's new

Auto-review is the rebranded, stabilised version of what Codex used to ship as Guardian / approvals reviewer behind a feature flag. The key pull requests that flipped it stable:

  • #19063 chore(auto-review) feature => stable — the flag flip itself.
  • #18021 Guardian → Auto-Review rename.
  • #18504 + #19056 — config + variant rebrand to auto-review.
  • #19113 — protocol/SDK config compatibility.
  • #18959 feat(auto-review) policy config — policy-driven evaluation.
  • #18890 feat(auto-review) short-circuit — skip unnecessary reviews when the answer is obvious.
  • #19050 feat(request-permissions) approve with strict review — a stricter review path for higher-risk requests.

The flow is simple: Codex is about to run a permissioned action → the reviewer agent evaluates the action in an isolated session (MCPs, plugins and hooks are explicitly disabled inside the review per #18722) → the app surfaces a risk level plus status. A Timed out review is treated as distinct from a denial and gets its own guidance (#17381, #17486, #17521, #17557).

Why it matters

Every serious Codex workflow in 2026 — parallel cloud tasks, overnight thread automations, GitHub PR babysitters, the new macOS Computer Use mode — fights the same boss: the approval dialog. One permission pop-up can freeze a 7-hour run for hours. Queue five agents in parallel and you get five streams of pop-ups competing for your attention.

Auto-review replaces that tax with a pre-judgement. Instead of agent → human decision, it becomes agent → reviewer agent → human override. The human keeps final authority, but only on the subset of actions the reviewer cannot confidently approve. That is the whole reason the tweet framed this as "longer, safer autonomous workflows with fewer approvals".

Technical facts

  • Stable version: codex-cli 0.124.0, released 2026-04-23 (same release as GPT-5.5 + Codex app updates).
  • Statuses surfaced in the app: Approved, Denied, Stopped, Timed out.
  • Isolation: MCP servers, plugins and hooks are disabled inside a review session to keep the reviewer from being influenced by external tools.
  • Efficiency: follow-up reviews now send only transcript deltas instead of the full history (#17269); stable review IDs across app-server events prevent lost or duplicated tracking (#17298).
  • Denied actions can be unstuck: #18955 added plumbing to manually approve stored Auto-Review denials.
  • Debug config: guardian_policy_config (#18923) — the rename to auto-review is cosmetic, the internal machinery still uses Guardian naming in places.
  • Analytics: dedicated guardian-review event schema, truncation metrics, TTFT plumbing (#17055, #17692, #17695, #17696).

The surrounding Codex numbers make Auto-review feel load-bearing: GPT-5.3-Codex hits 56.8% on SWE-Bench Pro, 77.3% on Terminal-Bench 2.0 (+13.3 pts over 5.2) and 64.7% on OSWorld-Verified (+26.5 pts), while running ~25% faster per task. More capability per turn means more turns, means more permission prompts — and Auto-review is the friction valve.

Comparison

The philosophical split with Claude Code is crisp:

ApproachClaude CodeCodex Auto-review
Default autonomy postureFull hand-off, --dangerously-skip-permissions / auto-approveHuman stays in loop, AI reviewer pre-judges each request
Who decides riskThe main agent, then the human post-hocA separate reviewer agent, before the action
Friction modelLow during task, high if it goes wrongModerate during task, lower blast radius
Best fitTrusted deep runs, big refactorsParallel queues, long-running autonomous loops, regulated environments

Before Auto-review, a comparable Codex session required either (a) clicking through every prompt or (b) opting into full-access modes that OpenAI itself is uncomfortable with — GPT-5.3-Codex is the first model they classify as "High" cybersecurity risk. Auto-review is OpenAI's attempt to give developers long-run autonomy without forcing them into the unsafe default.

Use cases

  • Overnight + scheduled threads. Thread automations wake the same thread on a schedule; Auto-review is what stops a run from stalling for 6 hours waiting on a single permission click.
  • Parallel cloud tasks and multi-agent PR runs. Developers report queueing 4-5 Codex tasks before morning coffee; with Auto-review every action is pre-screened concurrently, and self-reported success rates on "well-scoped maintenance work" jump from roughly 40-60% to 85-90%.
  • GitHub workflows. @codex mentions on PRs/Issues plus the bundled PR babysitter skill now have a real safety net — autonomous edits get a pre-flight risk read before they ship.
  • Computer Use on macOS. When Codex is driving Figma, Xcode, Slack or the browser, the attack surface expands to the whole desktop. Pre-screening every click and keystroke is no longer optional.
  • Teams with policy requirements. guardian_policy_config plus approve with strict review let teams standardise what counts as low/high risk across shared worktrees.

Limitations & pricing

  • Still failing half of professional tasks. 56.8% SWE-Bench Pro is best-in-class but also the same reality check as every prior model. Auto-review reduces click-tax, not reviewer responsibility.
  • Reviewer can time out. A timeout is not an approval — it is its own status with its own UX path. Plan for it.
  • MCPs, plugins and hooks are muted inside reviews. Intentional for isolation, but it means custom context your team relies on will not be visible to the reviewer agent.
  • Computer Use limits. macOS-only at launch; not available in the EU, UK or Switzerland. The Codex app itself runs on Windows since 2026-03-04.
  • No GPT-5.3-Codex API yet. OpenAI delayed API access because of the High cybersecurity classification. API users stay on gpt-5.2-codex ($1.25 / 1M input, $10 / 1M output).
  • Plans: bundled with ChatGPT — Plus $20/mo (30-150 msgs / 5hr window), Pro $200/mo (300-1,500 msgs), Business $25-30/user/mo, plus a new $100 Codex add-on (2× limits + GPT-5.3-Codex-Spark) live through at least 2026-05-31. Billing is token-based since 2026-04-02; agentic and image-gen tasks burn 3-5× faster than plain text.

What's next

Auto-review is the foundation, not the destination. What is signalled next:

  • Full (non-localhost) in-app browser autonomy — the last gap between Codex and Claude Code's computer-use story.
  • EU / UK / Switzerland rollout for Computer Use (no date committed).
  • GPT-5.3-Codex API availability "once it's safely enabled."
  • Enterprise agent internet access (still "coming soon").
  • Agent-to-agent delegation — the real multi-agent frontier, where Auto-review policies will likely carry across delegated subtasks.

The meta-story: as coding agents gain desktop control and long-horizon autonomy, the bottleneck moves from capability to governance. Auto-review is the first shipping answer to "how does one human supervise five agents running for seven hours each" — and it will get copied fast.

Nguồn: developers.openai.com/codex/changelog, BuildFastWithAI, AI Tool Analysis, Remio.