Seedance 2.0 API — Coming SoonGet early access
GPT-5.4 Release Date (2026): Latest News, Leaked Features & Developer Guide
guide

GPT-5.4 Release Date (2026): Latest News, Leaked Features & Developer Guide

EvoLink Team
EvoLink Team
Product Team
March 5, 2026
6 min read

If you're tracking GPT-5.4, you're probably trying to answer one question: should you wait for it, or build with what's available now?

This page separates confirmed signals, credible reporting, and speculation so you can make that decision quickly.

Update (March 5, 2026): GPT-5.4 is now listed on OpenRouter (openai/gpt-5.4) with posted pricing ($2.50 / 1M input, $0.625 / 1M cached input, $20.00 / 1M output), 1M context, and 128K max output. OpenAI direct billing tiers and enterprise contract pricing can still differ by channel.

Timeline So Far

Here are the most credible signals in chronological order:

  • February 27, 2026: Codex PR #13050 added original-resolution image support, with the minimum model version initially set to GPT-5.4. After seven force pushes within five hours, the threshold was changed to GPT-5.3-Codex. The PR was merged on March 3. (Source: GitHub PR #13050)
  • March 2, 2026: Codex PR #13212 added a /fast slash command, originally described as "toggle Fast mode for GPT-5.4." The reference was scrubbed within three hours. (Source: Awesome Agents)
  • March 2, 2026: Separately, OpenAI Codex team member Tibo accidentally posted a screenshot on X showing GPT-5.4 as a selectable model in the Codex app alongside GPT-5.3-Codex. The post was quickly deleted. (Source: NxCode, eWeek)
  • March 3, 2026: OpenAI posted "5.4 sooner than you think" on X.
  • March 3, 2026: alpha-gpt-5.4 briefly appeared in a public API models endpoint before being removed.
  • March 4, 2026: The Information reported GPT-5.4 may include a context window exceeding 1 million tokens and an "extreme" thinking mode.
  • March 4, 2026: PiunikaWeb reported GPT-5.4 activity on LMSYS Arena, suggesting internal testing.
  • March 5, 2026: OpenRouter listed openai/gpt-5.4 with public token pricing and limits.

Confirmed vs Speculative

TopicCan reasonably citeStill uncertainWhy it matters
AvailabilityOpenRouter now lists openai/gpt-5.4 (March 5, 2026)OpenAI direct tier parity and contract-tier differencesRollout and procurement decisions
Context windowOpenRouter listing shows 1M contextCross-provider parity and practical quality at full lengthLong-context architecture decisions
Reasoning mode"Extreme" mode is still mostly report-basedPublic mode controls, latency tiers, and defaultsResearch and analysis workloads
Vision detailLeak hints at full-resolution optionsActual quality and supported formatsImage analysis pipeline planning
Agentic improvementsMultiple hints from code referencesScope of tool-calling/agent upgradesMigration effort for agent flows
PricingOpenRouter lists $2.50 in / $20 out (+ cached input)OpenAI direct and enterprise pricing detailsBudget forecasting

Should You Wait or Build Now?

Should you wait for GPT-5.4 or build now on GPT-5.2

Build now on GPT-5.2 if:

  • Your product ships in the next 2 weeks.
  • You do not need more than 400K context.
  • You are latency-sensitive.
  • You are already in production and only need a model swap later.

Use GPT-5.4 now (in controlled rollout) if:

  • Your design depends on 1M context right now.
  • Your team can run side-by-side evals on quality, latency, and cost.
  • You already have model routing and fallback in place.
  • You can accept provider-level variance during early adoption.

Recommended approach: keep GPT-5.2 as baseline, route a limited share to GPT-5.4, then promote only after your eval gates pass.

How to Prepare During Early Rollout

1. Set up model-agnostic routing

Keep one internal inference interface and route models behind it. This turns future migration into a config update.

2. Build an eval suite now

Test against your real failure modes:

  • Your hardest real task
  • One long-context scenario
  • One regression set for simple tasks
  • One cost check (tokens and dollars per task)

3. Define success criteria in advance

Pick a few product-level metrics before increasing GPT-5.4 traffic:

  • Task completion quality
  • P95 latency
  • Cost per task
  • Hallucination rate for your domain

With GPT-5.4 now publicly listed via OpenRouter, EvoLink integration planning should prioritize baseline checks:

  • Availability and stability under load
  • Latency baseline (P50/P95)
  • Error handling behavior
  • Quality gate against current GPT-5.2 evals
Current GPT-5.2 pricing on EvoLink is listed as $1.40/1M input and $11.20/1M output. GPT-5.4 EvoLink final pricing should be confirmed on EvoLink pricing pages at rollout time.

GPT-5 Family Snapshot

ModelDateContextPositioningEvoLink price
GPT-5.3 InstantMarch 3, 2026128K (API alias: gpt-5.3-chat-latest)Everyday tasksN/A
GPT-5.2 ThinkingDecember 11, 2025400KDeeper reasoning$1.40/1M input
GPT-5.2-CodexDecember 18, 2025 (OpenAI release) / January 14, 2026 (Copilot GA)400KAgentic coding$1.40/1M input
GPT-5.1November 2025400KGeneral-purpose$1.00/1M input
GPT-5.4March 2026 (listed on OpenRouter)1M (OpenRouter listing)Flagship upgradeTBD on EvoLink

FAQ

When will GPT-5.4 be released?

GPT-5.4 is already listed on OpenRouter as of March 5, 2026. OpenAI direct-channel rollout details can still vary by tier.

Is GPT-5.4 available in OpenAI API right now?

It is listed on OpenRouter (openai/gpt-5.4). Direct OpenAI API availability and pricing details may differ by account tier and contract.

Will GPT-5.4 be more expensive than GPT-5.2?

On current OpenRouter listing, yes: GPT-5.4 is priced above GPT-5.2. Validate your effective cost with your own prompt mix and cache hit rate.

Will GPT-5.4 replace GPT-5.2 immediately?

Probably not. Major releases usually overlap with prior models for a transition period.

Will existing OpenAI-compatible code keep working?

Most likely yes. In most integrations, migration should primarily be a model-name change plus evaluation checks.

What is the safest GPT-5.2 to GPT-5.4 migration path for production?

Use model-agnostic routing, keep feature flags per model, run domain evals, then do staged rollout by traffic percentage.

Will gpt-5.3-chat-latest automatically become GPT-5.4?

Do not assume that. Treat aliases as separate products and pin explicit model IDs in production.

Does GPT-5.4 help long-context RAG quality or only increase token limit?

Higher context can help only if retrieval quality, chunking strategy, and evaluation coverage are already strong.

Should startups ship with GPT-5.2 now or wait for GPT-5.4 in March 2026?

If launch is near-term, ship on GPT-5.2 and prepare a fast model-switch path. Waiting is mainly justified for hard 1M+ context requirements.

How will GPT-5.4 compare with Gemini 3.1 Pro and Claude Opus 4.6?

A fair comparison requires production access and side-by-side tests on identical tasks.


Last checked: March 5, 2026

Ready to Reduce Your AI Costs by 89%?

Start using EvoLink today and experience the power of intelligent API routing.