AI Models
OpenAI Launches GPT-5.4 Mini and Nano, Its Most Capable Small Models for the Agentic Era
Mini runs 2x faster than GPT-5 mini while approaching full GPT-5.4 performance; Nano costs just $0.20 per million input tokens, targeting subagent and classification workloads.
OpenAI on Tuesday released GPT-5.4 Mini and GPT-5.4 Nano, two new models that represent the company’s most aggressive push yet into the economics of the agentic era. Mini, the larger of the pair, runs twice as fast as the previous GPT-5 mini while approaching full GPT-5.4 performance on rigorous benchmarks including SWE-Bench Pro and OSWorld-Verified — the kind of evaluations that measure whether a model can actually write production code and navigate real operating system tasks rather than ace multiple-choice trivia. OpenAI is positioning Mini as the default workhorse for its API, Codex coding assistant, and the ChatGPT consumer product, where latency and throughput matter as much as raw capability.
Nano, meanwhile, is designed for the unglamorous but economically enormous layer of AI infrastructure that sits beneath the models users interact with directly. At $0.20 per million input tokens — a price point that would have seemed absurd eighteen months ago — Nano targets classification, extraction, routing, and subagent tasks: the thousands of small decisions that an orchestration framework makes for every user-visible action. In a world where a single agentic workflow might chain a dozen model calls to research a topic, draft a document, and verify its claims, the cost of each individual call becomes the binding constraint on what’s economically viable to automate.
The strategic implications extend well beyond OpenAI’s own products. As multi-agent architectures proliferate — frameworks like LangGraph, CrewAI, and Anthropic’s own orchestration tools are now standard in enterprise deployments — the market is bifurcating between “thinking” models that reason carefully about hard problems and “doing” models that execute routine subtasks at machine speed. Mini and Nano are OpenAI’s bid to own the doing layer, the high-volume substrate on which every agentic application will run. With Mini available immediately across the API, Codex, and ChatGPT, and Nano as an API-only offering, OpenAI is signaling that the next phase of the AI cost curve will be measured not in dollars per frontier query but in fractions of a cent per agent action.