Seedance 2.0 API — Coming SoonGet early access

GPT-5.4 API

GPT-5.4 API on EvoLink provides access to OpenAI's most capable frontier model through a single endpoint. The first general-purpose model with native computer use, 1.05M-token context, 128K max output, and significantly better token efficiency than GPT-5.2, meaning faster responses and lower effective cost.
Run With API
Using coding CLIs? Run GPT-5.4 via EvoCode — One API for Code Agents & CLIs. (View Docs)

GPT-5.4 API

The first general-purpose model with native computer use, 1.05M-token context, 128K max output, and significantly improved token efficiency.

GPT-5.4 API

Billing Rules

  • Input/output billed per 1M tokens.
  • Cached input: 90% discount.
  • >272K input: 2x input + 1.5x output for full session.
  • Regional processing: 10% uplift.
  • Reasoning tokens count as output.
$

PRICING

PLANCONTEXT WINDOWMAX OUTPUTINPUTOUTPUTCACHE READ
GPT-5.41.05M128.0K
$2.00-20%
$2.50Official Price
$12.00-20%
$15.00Official Price
$0.200-20%
$0.250Official Price
GPT-5.4 (Beta)1.05M128.0K
$0.650-74%
$2.50Official Price
$3.90-74%
$15.00Official Price
$0.065-74%
$0.250Official Price

Pricing Note: Price unit: USD / 1M tokens

Cache Hit: Price applies to cached prompt tokens.

Two ways to run GPT-5.4 — pick the tier that matches your workload.

  • · GPT-5.4: the default tier for production reliability and predictable availability.
  • · GPT-5.4 (Beta): a lower-cost tier with best-effort availability; retries recommended for retry-tolerant workloads.
ModelMetricOfficialEvoLinkCredits
GPT-5.4Input (Standard)$2.50 / 1MTBATBA
GPT-5.4Input (Cached)$0.25 / 1MTBATBA
GPT-5.4Input (>272K Prompt Tier)$5.00 / 1MTBATBA
GPT-5.4Output (Standard)$15.00 / 1MTBATBA
GPT-5.4Output (>272K Prompt Tier)$22.50 / 1MTBATBA

If it is down, we automatically use the next cheapest available option, ensuring 99.9% uptime at the best possible price.

Capabilities

Native computer use: first general-purpose model that operates computers

GPT-5.4 is the first general-purpose model with native, state-of-the-art computer-use capabilities. It can click, type, and navigate software with screenshots plus keyboard/mouse commands without requiring a separate specialized model. On OSWorld-Verified, GPT-5.4 scores 75.0%, surpassing human performance at 72.4%.

GPT-5.4 API showcase of AI platform feature 2

1.05M-token context with better token efficiency

Process entire codebases, book-length documents, or months of conversation history in one request. GPT-5.4vides 2.6x GPT-5.2's 400K context and uses significantly fewer tokens for equivalent tasks, reducing usage and improving speed.

GPT-5.4 API showcase of AI platform feature 3

Deep reasoning with adjustable effort

Use five reasoning levels: none, low, medium, high, and xhigh. For the hardest tasks, increase effort to deepen reasoning quality. GDPval reaches 83.0% (matching or exceeding professionals across 44 occupations) versus GPT-5.2 at 70.9%.

GPT-5.4 API showcase of AI platform feature 4

Why Developers Choose

Frontier capability, broader tools, and practical integration through EvoLink.

Full tool ecosystem with Tool Search

Web search, file search, image generation, code interpreter, hosted shell, computer use, MCP, and tool search are natively supported. Tool Search helps agents select and use the right tools across large connector ecosystems.

Better results with fewer tokens

GPT-5.4 is OpenAI's most token-efficient reasoning model. Compared with GPT-5.2, it generally uses fewer tokens for equivalent tasks, often improving speed and effective cost per job.

One key, zero setup

Access GPT-5.4 with one EvoLink API key. Migration from GPT-5.2 is drop-in for most integrations by changing one model string.

How to Integrate

Three steps from key creation to production monitoring.

1

Get your API key

Sign up on EvoLink, generate your API key, and use it immediately with GPT-5.4 and 47+ other models.

2

Send your request

POST with model set to "gpt-5.4", your messages array, and optional parameters.

3

Deploy and monitor

Track usage, costs, and reasoning-token consumption in the EvoLink dashboard and scale workflows when ready.

Key Features

Core strengths for production agents, coding systems, and enterprise workflows.

Context

1.05M Context Window

Process entire repositories and book-length documents in one request.

Output

128K Max Output

Generate complete reports and long implementations in one response.

Computer Use

Native Computer Use

Operate software via screenshots and keyboard/mouse commands (OSWorld 75.0%, human 72.4%).

Tooling

Tool Search

Agents can automatically identify and use the right tools in larger ecosystems.

Efficiency

Token Efficiency

Uses fewer tokens than GPT-5.2 for equivalent problem-solving in many workloads.

Cost

Prompt Caching

Cached input pricing at $0.25 per 1M tokens, a 90% discount from standard input.

Benchmarks: GPT-5.4 vs GPT-5.2

Verified benchmark deltas highlight stronger professional performance, tool use, browsing quality, and computer-use reliability.

BenchmarkGPT-5.4GPT-5.2
GDPval83.0%70.9%
SWE-Bench Pro57.7%55.6%
OSWorld (Human: 72.4%)75.0%47.3%
Toolathlon54.6%46.3%
BrowseComp82.7%65.8%
MMMU-Pro81.2%79.5%
Factual errors per claim33% fewerBaseline
Factual errors per response18% fewerBaseline

Data Summary

GPT-5.4

gpt-5.4-2026-03-05 | $2.50/$15/$0.25 | 1.05M/128K | reasoning none→xhigh | all tools

GPT-5.4 Thinking

ChatGPT only, not an API model

What Changed from V1

  • Added GPT-5.4 Thinking clarification (ChatGPT only, not an API model).
  • Moved native computer use to lead capability (OSWorld 75.0% > human 72.4%).
  • Added token-efficiency positioning (fewer tokens, lower effective cost).
  • Added Tool Search capability details.
  • Added benchmark comparison section versus GPT-5.2.
  • Updated SEO title and meta description for quick-start intent.

Frequently Asked Questions

Everything you need to know about the product and billing.

GPT-5.4 pricing is $2.50 per 1M input, $15.00 per 1M output, and $0.25 per 1M cached input. For prompts above 272K input tokens, the full session is billed at 2x input and 1.5x output.
GPT-5.4 is the API model on this page. GPT-5.4 Thinking is a ChatGPT product mode, not a separate API model.
GPT-5.4 is priced at $2.50/$15 with 1.05M context and native computer use. Claude Opus 4.6 is $5/$25 up to 200K and $10/$37.50 above 200K. Gemini 3.1 Pro is $2/$12 up to 200K and $4/$18 above 200K. GPT-5.4 is the only one in this group with native computer use plus flat pricing up to 272K.
When input exceeds 272K tokens, the full session is billed at 2x input and 1.5x output. GPT-5.4 becomes $5.00 input and $22.50 output per 1M tokens.
Native computer use means the model can operate computers using screenshots plus keyboard and mouse actions. It can browse websites, navigate apps, and complete multi-step interface workflows. GPT-5.4 scores 75.0% on OSWorld-Verified, above the human baseline of 72.4%.
Yes. GPT-5.4 is more token-efficient on many difficult tasks. Although list price per token is higher than GPT-5.2, total token consumption is often lower, so effective cost per completed task can be similar or lower.
GPT-5.4 supports web search, file search, image generation, code interpreter, hosted shell, computer use, MCP, tool search, apply patch, and skills.

Related Resources

Internal links for release notes, pricing analysis, comparisons, and migration decisions.