OpenAI launches GPT-5.4 mini and nano AI models

AI News


OpenAI We have released two smaller versions, GPT-5.4 mini and GPT-5.4 nano. A.I. It’s a model designed for high-volume workloads as organizations aim to balance performance, cost, and latency in production environments.

Announced via a LinkedIn post from OpenAI for Business, the model extends the capabilities of GPT-5.4 into a more efficient format, supporting use cases such as coding assistants, subagents, and real-time multimodal applications. This release reflects a broader shift in AI adoption, where speed and responsiveness are now as important as raw model capabilities.

“Today, we are releasing GPT-5.4 mini and GPT-5.4 nano, our most powerful compact models to date,” the company said in a LinkedIn post.

He added, “It brings many of the strengths of GPT-5.4 into a faster, more efficient model built for high-volume workloads.”

Improving performance with a focus on speed and cost efficiency

GPT-5.4 mini improves on GPT-5 mini for coding, reasoning, multimodal understanding, and tool usage, making it run more than twice as fast. Benchmarks show it approaches the performance of the larger GPT-5.4 model in ratings such as SWE-Bench Pro and OSWorld-Verified.

GPT-5.4 nano is positioned as the smallest and lowest-cost model in the series, designed for tasks such as classification, data extraction, ranking, and supporting coding workflows through subagents.

The company says, “GPT-5.4 mini improves on GPT-5 mini in all aspects of coding, reasoning, multimodal understanding, and tool usage, making it run more than twice as fast.”

“GPT-5.4 nano is our smallest and cheapest GPT-5.4 model optimized for classification, data extraction, ranking, and subagent coding.”

This model is designed for environments where latency directly impacts the user experience, such as systems that require rapid iteration or real-time interaction.

Migration to multi-model systems and subagents

OpenAI positions new models as part of a broader architecture where multiple models work together. In this setup, larger models handle planning and decision-making, while smaller models perform specific tasks quickly at scale.

This approach is particularly relevant for coding workflows and enterprise AI systems that require varying levels of reasoning between tasks. Smaller models, such as the GPT-5.4 mini, can handle subtasks like navigating the codebase and processing documents, while larger models manage coordination and final output.

The company said in a post on LinkedIn: “These models are designed for responsive coding assistants, subagents, and multimodal applications that require low latency without sacrificing strong performance.”

Availability across API, Codex, and ChatGPT

GPT-5.4 mini is available across OpenAI’s API, Codex, and ChatGPT and can be used for inputting text and images, using tools, and computer-based workflows. GPT-5.4 nano is currently available through the API.

The pricing structure is geared toward scalability, with a lower cost per token compared to larger models, making it suitable for applications that require frequent or continuous usage.

This release highlights changes in the way AI is being deployed across sectors, including education and EdTech. Rather than relying solely on large, general-purpose models, more and more organizations are building systems that combine models of various sizes to optimize both performance and cost.

ETIH Innovation Award 2026

ETIH Innovation Award 2026 is now open and highly appreciatedStrengthen your educational technology organization to deliver measurable impact across K-12, higher education, and lifelong learning. The award welcomes applications from the UK, the Americas and overseas, and applications will be assessed on the basis of evidence of achievement and practical application.



Source link