OpenAI Launches High-Speed GPT-5.4 Mini and Nano Models
- •OpenAI releases GPT-5.4 mini and nano models optimized for high-speed coding and autonomous subagents
- •GPT-5.4 mini achieves 54.4% on SWE-Bench Pro, outperforming previous iterations while running twice as fast
- •GPT-5.4 nano offers cost-effective intelligence at $0.20 per million tokens for lightweight classification tasks
OpenAI has expanded its flagship lineup with the debut of GPT-5.4 mini and nano, two lightweight models engineered to balance sophisticated reasoning with rapid execution. These models are designed for high-volume workloads where the speed of a response (latency) is just as critical as its accuracy. While the flagship GPT-5.4 handles complex logic, these smaller siblings excel at powering responsive coding assistants and navigating digital user interfaces.
The technical leap is most evident in the mini version, which delivers over twice the speed of its predecessor. In rigorous evaluations like SWE-Bench Pro—a test measuring an AI's ability to resolve real-world software issues—the mini model nearly matched the performance of the full-scale GPT-5.4. This efficiency allows developers to build systems where a large model acts as a supervisor that delegates specific, narrow tasks to faster subagents running in parallel.
Accessibility remains a core focus, with the nano model priced for high-frequency, low-complexity operations like data extraction or basic debugging. By integrating these models into the Codex platform and ChatGPT, OpenAI is pushing toward a future where AI isn't just one giant brain, but a coordinated network of specialized models working at different scales.