OpenAI changed Codex pricing on April 3. Pay-as-you-go token billing, business seats dropped to $20/mo, and they claim Codex uses 4x fewer tokens than Claude Code. The claim holds up in benchmarks: 1.5M tokens vs 6.2M for the same task. But Claude's extra tokens buy you more thorough code with fewer errors. Codex wins Terminal-Bench at 77.3% vs 65.4%. Claude Code wins on code quality, 1M context window, and cross-platform support. The smart move: use both. Codex for speed and cost-sensitive work. Claude Code for architecture and complex features.
- Codex switched to pay-as-you-go token billing on April 3, 2026. Business seats dropped from $25 to $20/mo (OpenAI blog)
- OpenAI claims Codex CLI uses 4x fewer tokens than Claude Code for equivalent tasks
- Figma-to-code benchmark: Codex used 1.5M tokens vs Claude Code's 6.2M for comparable output (Morphllm)
- Terminal-Bench 2.0: Codex GPT-5.3 scores 77.3% vs Claude Code at 65.4%
- Claude Code's 1M token context window is 4x larger than Codex's 256K
- Codex app supports macOS and Windows as of March 2026. Codex CLI also runs on Linux. No self-hosting option.
- Codex usage has grown 6x since January 2026, with 2M+ developers using it weekly (OpenAI)
- Codex-only seats get up to $500 in promotional credits per member
OpenAI made two moves in one week. On April 2, they killed Sora. On April 3, they changed how Codex charges you. The message is clear: coding agents are where OpenAI is putting its chips.
The headline claim: Codex uses 4x fewer tokens than Claude Code. If true, that means the same work costs a quarter of the price. That's not a small difference.
I looked at the benchmarks, the actual token numbers, and what developers who use both are saying. The 4x claim is real. The conclusion it implies is not.
What Codex Changed on April 3
Pay-as-you-go replaces fixed seats
Before April 3, Codex came bundled with ChatGPT subscriptions or as fixed-price team seats. Now OpenAI offers two options for business customers:
Standard ChatGPT Business seats (dropped from $25 to $20/mo) include Codex with a usage cap. Codex-only seats bill purely on token consumption with no rate limits. Eligible workspaces get up to $500 in promotional credits per Codex member, which is enough to test it seriously before committing.
For individual developers, nothing changes on the surface: ChatGPT Plus at $20/mo includes Codex. But the usage caps on Plus matter. One developer on the OpenAI forum reported credits depleting within hours during the March 28-30 period, and Codex limits resetting on April 1 was described as "finally good."
Claude Code pricing stays the same: $20/mo (Pro) or $100-200/mo (Max). No usage-based billing. No surprises. You hit a limit, you wait. Compare both in our AI cost calculator.
The 4x Token Efficiency Claim
It's real, but it doesn't mean what you think
In a Figma-to-code benchmark (cited by Morphllm and Builder.io), Codex CLI completed the task using 1.5 million tokens. Claude Code used 6.2 million tokens for comparable output. That's a genuine 4x gap.
On API pricing, that translates to roughly $15 per complex task with Codex versus $155 with Claude Code. Ten times the cost for the same deliverable. At first glance, Claude Code looks absurd.
But there's a reason Claude uses more tokens. It generates more thorough output. More comments, more error handling, more edge cases covered on the first pass. Developers who use both report Claude's code needs less rework. Codex is faster and cheaper per task, but you might end up running it twice to catch what Claude caught the first time.
The 4x efficiency claim is about tokens consumed, not about value delivered per token. Those are different questions.
The analogy
Codex writes a fast first draft. Claude Code writes a more careful first draft. The fast draft costs less. The careful draft needs less editing. Which is cheaper depends on how much your editing time costs.
Benchmark Numbers
Where each one wins
Head-to-head benchmarks
| Benchmark | Codex (GPT-5.3) | Claude Code (Opus 4.6) | Winner |
|---|---|---|---|
| Terminal-Bench 2.0 | 77.3% | 65.4% | Codex (+12 pts) |
| SWE-bench Verified | ~80% | 80.8% | Claude Code (marginal) |
| Token efficiency (Figma task) | 1.5M tokens | 6.2M tokens | Codex (4x fewer) |
| Context window | 256K tokens | 1M tokens | Claude Code (4x larger) |
| Platform support | macOS, Windows, CLI on Linux | macOS, Linux, Windows (WSL) | Tie |
Not sure which AI model to use?
12 models · Personalized picks · 60 seconds
Codex is better at terminal-native tasks (scripts, DevOps, CLI tools) by a significant margin. Claude Code is better at complex code that requires understanding large codebases, because it can hold 4x more context in memory. SWE-bench (real-world coding tasks) is basically a tie.
What Each One Feels Like
Two different philosophies
Codex runs both locally and in the cloud. You can kick off a task and it runs asynchronously while you do other things. It picks which model handles your task internally based on complexity. Some developers find this convenient. Others find it annoying because you can't choose the model yourself.
Claude Code is local and synchronous by default. It runs in your terminal, against your codebase, and you watch it work. The new auto mode reduces permission interrupts, and Channels lets you message it from Telegram, but it's fundamentally a tool that runs where you are.
One practical note: the Codex desktop app launched on macOS first and added Windows in March 2026. The Codex CLI also runs on Linux. Claude Code runs everywhere (macOS, Linux, Windows via WSL). Both tools now have broad platform support, though Codex's desktop app is still newer on Windows.
Platform note
Codex app is on macOS and Windows (as of March 4, 2026). Codex CLI also runs on Linux. Claude Code runs on macOS, Linux, and Windows via WSL. Both cover most setups now. Codex still has no self-hosted or on-premises option.
Real Monthly Cost
Same $20 sticker price, very different actual bills
What you'll actually pay
| Usage | Claude Code | Codex |
|---|---|---|
| Light (10-20 tasks/mo) | $20 (Pro covers it) | $20 (Plus covers it) |
| Medium (50 tasks/mo) | $20 (Pro, may hit limits) | $20 (Plus, may hit limits) |
| Heavy (100+ tasks/mo) | $100-200 (Max) | $20 + pay-as-you-go overages |
| API cost per complex task | ~$155 (6.2M tokens) | ~$15 (1.5M tokens) |
| Surprise charges possible? | No (flat pricing) | Yes (pay-as-you-go seats) |
For light to medium use, both cost $20/mo and the difference is negligible. For heavy use, Claude Code is predictable but expensive (Max at $100-200). Codex is cheaper per task but unpredictable on the pay-as-you-go plan. If you've seen what happened with Cursor's overage pricing, you know why unpredictable billing makes developers nervous.
What Developers Complain About
Real issues from real users
Codex complaints (from OpenAI forums and Hacker News):
- Credits depleted within hours during late March, limit resets described as "finally good" on April 1
- Can't choose which model handles your task. Codex picks internally. Annoying if you know which model you want.
- Was macOS only until March 2026. Windows added, CLI runs on Linux, but still no self-hosting.
- Environment setup is difficult. You can't spin up containers needed for tests, limiting usefulness on complex projects.
- All tasks run on OpenAI's cloud. No on-premises option. Defense, banking, and healthcare teams can't use it.
Claude Code complaints (from Reddit and Hacker News):
- Rate limits hit too quickly on Pro. One developer lost an entire afternoon waiting 5 hours for a reset mid-debugging.
- Uses 3-4x more tokens than Codex for the same task. Expensive on API pricing.
- No model choice either, but at least you know it's always Claude.
Who Wins Where
Different tools for different jobs
Winner by task
| Task | Winner | Why |
|---|---|---|
| DevOps and scripts | Codex | Terminal-Bench 77.3%, optimized for CLI workflows |
| Large codebase refactoring | Claude Code | 1M context window holds entire repos in memory |
| Cost-sensitive batch work | Codex | 4x fewer tokens per task |
| Code quality (first draft) | Claude Code | More thorough output, fewer errors on first pass |
| Self-hosting / on-premises | Claude Code | Local execution possible. Codex is cloud-only. |
| Async background tasks | Codex | Cloud-based execution, works while you don't |
| Privacy-sensitive orgs | Claude Code | Local execution possible. Codex is cloud-only. |
| MCP integrations | Claude Code | Full MCP support for external tools. Codex has limited. |
The Verdict
Cheaper isn't always better
The decision
- 1Terminal-heavy workflow (DevOps, scripts, CLI)? Codex is measurably better.
- 2Working on large codebases that need full context? Claude Code's 1M window wins.
- 3On a tight budget for batch coding tasks? Codex at 4x fewer tokens saves real money.
- 4Need self-hosted or on-premises? Claude Code. Codex is cloud-only.
- 5Need on-premises or self-hosted? Claude Code. Codex is cloud-only.
- 6Want both strengths? Use both. Codex for speed tasks, Claude Code for quality tasks.
The 4x token efficiency is real and it matters for cost. If you're running 100 coding tasks a month through API pricing, Codex saves you hundreds of dollars. But cheaper per token doesn't mean better per task. Claude Code uses more tokens because it does more work per generation. The code it writes tends to be more complete and needs less fixing.
The practical answer for most developers: try Codex on the $500 promotional credit. See how it handles your actual workflow. Compare the output quality to what Claude Code gives you. If Codex's faster, lighter output works for your tasks, you'll save money. If you find yourself re-running tasks to get complete output, Claude Code's extra tokens are buying you something.
I keep coming back to the same pattern across every tool comparison I've written: the best developers use two tools, not one. Codex for the fast stuff. Claude Code for the careful stuff. $40/mo total and you get the strengths of both.
FAQ
Is Codex really 4x more token-efficient than Claude Code?
In a Figma-to-code benchmark, yes. Codex used 1.5M tokens vs Claude Code's 6.2M for comparable output. But Claude's extra tokens produce more thorough code with better error handling. Efficiency in tokens and efficiency in outcome are different things.
How much does Codex cost after the April 2026 change?
ChatGPT Plus ($20/mo) includes Codex with usage caps. Business seats dropped from $25 to $20/mo. Codex-only seats use pay-as-you-go token billing with no rate limits. New workspaces get up to $500 in promo credits per member.
Which is better for coding?
Codex wins on terminal tasks (77.3% vs 65.4% on Terminal-Bench) and costs 4x less per task. Claude Code wins on complex refactoring (1M context window) and code quality (more thorough first drafts). Most experienced developers use both.
Does Codex work on Windows?
Yes, as of March 4, 2026. The desktop app runs on macOS and Windows. The Codex CLI also supports Linux. No self-hosted or on-premises option though - all tasks run on OpenAI's cloud.
Keep Reading
Stay ahead of the AI curve
We test new AI tools every week and share honest results. Join our newsletter.
![Claude Code vs OpenAI Codex: $20/mo Each but OpenAI Claims 4x Better Token Efficiency [2026] - Featured Image](/_next/image?url=%2Fimages%2Fclaude-code-vs-codex-2026.png&w=3840&q=75)


![Best Free AI Certifications Ranked [2026]](/_next/image?url=%2Fimages%2Fbest-ai-certifications-ranked-2026.png&w=3840&q=75)