OpenAI has released two new models — GPT-5.4 mini and GPT-5.4 nano — designed for speed, cost efficiency, and multi-agent workflows. GPT-5.4 mini is now live for free ChatGPT users, marking one of the most significant expansions of flagship-tier AI to the general public so far.

What’s New

Both models are compact versions of the full GPT-5.4, released two weeks ago. They’re not stripped-down fallbacks — they’re purpose-built for workloads where latency matters more than raw reasoning depth.

  • GPT-5.4 mini: More than 2x faster than GPT-5 mini, with major improvements across coding, reasoning, multimodal understanding, and tool use. Approaches GPT-5.4 performance on SWE-Bench Pro and OSWorld-Verified benchmarks. Available in ChatGPT Free, Go, and Pro tiers, plus the API and Codex.
  • GPT-5.4 nano: The smallest and cheapest model in the lineup, priced at roughly $0.20 per million input tokens. API-only for now. Built for classification, data extraction, ranking, and lightweight subagent tasks.

Why This Release Is Different

The framing around these models signals a shift in how OpenAI thinks about AI architecture. Rather than a single model handling everything, GPT-5.4 mini and nano are explicitly designed to work inside multi-agent systems — where a larger model plans and orchestrates, and smaller subagents execute specific tasks in parallel at speed.

In OpenAI’s own Codex platform, for example, GPT-5.4 handles planning and final judgment while GPT-5.4 mini subagents handle codebase search, large file review, and document processing simultaneously. This pattern is becoming the standard architecture for production AI systems.

What It Means for Developers

For developers building on the API, the cost difference is significant. The full GPT-5.4 costs $2.50 per million input tokens. GPT-5.4 nano comes in at $0.20 — a 12x reduction. For high-volume pipelines where speed and cost matter, that gap changes what’s economically viable to build.

GPT-5.4 mini is also now rolling out in GitHub Copilot across VS Code, JetBrains, Visual Studio, and Xcode — available to Pro, Pro+, Business, and Enterprise users via the model picker.

Performance on Coding Tasks

On GPQA Diamond (a test of expert-level reasoning), GPT-5.4 mini scored 88.01% — trailing the full GPT-5.4’s 93% by a relatively small margin. On coding benchmarks, it approaches flagship-level pass rates while running significantly faster. Notion’s engineering team reported that it often matches or beats more expensive models on complex formatting tasks at a fraction of the compute cost.

Conclusion

GPT-5.4 mini and nano aren’t just cheaper versions of GPT-5.4 — they’re the foundation of a new way of building AI systems, where speed and cost efficiency at the subagent level matter as much as raw intelligence at the top. If you’re building with AI or just using ChatGPT on the free tier, the experience just got meaningfully faster. Browse our AI coding tools directory to see how GPT-5.4 fits into the wider ecosystem.