OpenAI Releases o3 & o4-mini with Visual Reasoning

Just two days after unveiling GPT-4.1, OpenAI has dropped a surprise: two entirely new AI models, o3 and o4-mini — and they might just change the game.
Introducing o3: The Pinnacle of AI Reasoning
OpenAI describes o3 as its most advanced reasoning model yet. It shines in coding, math, and science, pushing the boundaries of what we’ve come to expect from large language models. But what truly sets o3 apart isn’t just raw power — it’s versatility.
For the first time, OpenAI’s reasoning models can use and combine all of ChatGPT’s tools — including web browsing, image generation, file handling, and code execution — in a single conversation. This means o3 can now tackle complex, multi-step problems independently, bridging gaps between modalities and enabling near-autonomous workflows.
Meet o4-mini: Lean, Affordable, Still Powerful
The second model, o4-mini, offers a cost-efficient alternative with surprisingly robust performance. It handles the same kinds of tasks as o3 — including reasoning, coding, and visual understanding — but in a more lightweight package. It’s designed for users who want intelligent output without the heavier compute price tag.
Visual Reasoning Enters a New Era
Both o3 and o4-mini represent a major leap in visual understanding. Not only can they “see” images, they can also interpret and reason about them on a deeper level. Upload a messy whiteboard photo, a complex sketch, or a hand-drawn diagram — these models will understand it, contextualize it, and even revise it as part of a reasoning process. It’s the kind of image-to-intelligence pipeline that previous generations could only partially emulate.
Tool Integration Unlocks Independent Problem Solving
This deep integration of tools — long a dream of AI researchers — is now a practical reality. By combining reasoning capabilities with tools like DALL·E, the browser, and Python, these models don’t just respond — they act. The result is more intelligent behavior across academic benchmarks, business use cases, and creative tasks.
Codex CLI: A Minimal Coding Companion
Alongside the new models, OpenAI introduced Codex CLI, a streamlined coding agent for developers. It connects local code environments with OpenAI’s models, giving devs a fast, focused interface to generate, test, and iterate on code — currently compatible with o3 and o4-mini, and soon, GPT-4.1.
A Pivot in OpenAI’s Roadmap
Interestingly, this release marks a change in OpenAI’s previous strategy. CEO Sam Altman had initially suggested o3 would not launch as a standalone product. But in early April, he reversed course, citing several reasons — including an opportunity to dramatically improve GPT-5, which is now expected in the coming months.
Altman also acknowledged the difficulty of integrating all capabilities into a single, seamless experience — an issue that pushed OpenAI to expand its current lineup before consolidating everything under GPT-5.
Available Now — With More to Come
As of today, ChatGPT Plus, Pro, and Team users can access o3 and o4-mini. A more powerful o3-pro version is also on the way, set to roll out to Pro users in the coming weeks. Until then, those users will continue using o1-pro.
With today’s release, OpenAI isn’t just pushing the envelope — it’s reshaping what intelligent systems can do in real-world applications. And this may just be the prelude to something even bigger.