ConversationTechSummitAsia

OpenAI Unveils o3 and o4-mini: Pioneering AI Reasoning Models to Elevate ChatGPT’s Capabilities

OpenAI, a trailblazer in the artificial intelligence sector, announced the release of its latest AI reasoning models, o3 and o4-mini, on Wednesday. These groundbreaking models are designed to enhance the performance of ChatGPT by incorporating advanced reasoning abilities that pause and analyze questions before responding. The introduction of these models signifies OpenAI’s strategic move to stay ahead in the fiercely competitive AI industry, taking on giants like Google, Meta, xAI, Anthropic, and DeepSeek.

### Revolutionary Advancements

The company has dubbed o3 as its most advanced reasoning model to date, with superior performance metrics across various domains including mathematics, coding, scientific reasoning, and visual comprehension. In addition, o4-mini is positioned as a cost-effective alternative, offering a balance between pricing, speed, and overall performance—key considerations for developers selecting AI models for their applications.

### Expanding ChatGPT’s Toolkit

Unlike its predecessors, o3 and o4-mini boast the ability to utilize ChatGPT’s available tools to generate responses, which includes:
– Web browsing
– Execution of Python code
– Image processing
– Image generation

Subscribers to OpenAI’s Pro, Plus, and Team plans can access these innovations starting today. Furthermore, a specialized version of o4-mini, known as “o4-mini-high,” is exclusively available, offering enhanced reliability by allowing more processing time for crafting responses.

### Keeping Ahead in the AI Race

The launch of these models comes amidst a high-stakes race in the AI technology realm. Although OpenAI was first to introduce an AI reasoning model, the competition quickly released their versions, matching and sometimes surpassing OpenAI’s offerings. This backdrop makes o3 and o4-mini pivotal in maintaining OpenAI’s edge.

Notably, o3’s release was momentarily uncertain. Initially, OpenAI CEO Sam Altman hinted at reallocating resources to develop a more sophisticated model incorporating o3’s technologies. But rising competitive pressures led to a swift decision to bring o3 to the market.

### Performance Credentials

OpenAI’s o3 achieves unprecedented performance with a score of 69.1% on the SWE-bench verification test—an assessment of coding proficiency. The o4-mini model claims similar results, with a 68.1% score, while a previous model, o3-mini, scored significantly lower at 49.3%.

### Thinking with Images

For the first time, OpenAI introduces models capable of ‘thinking with images.’ Users can upload visual content, such as diagrams or sketches, into ChatGPT for analysis during the model’s reasoning process. This novel capability allows o3 and o4-mini to interpret blurry or low-resolution images while performing image-related tasks like zooming or rotating.

### Additional Features

Besides image processing, o3 and o4-mini can execute Python code directly in the user’s browser, thanks to ChatGPT’s Canvas feature, and browse the web for real-time information. These models are also available for developers on OpenAI’s API platforms—Chat Completions and Responses APIs—enabling application development with flexible usage-based pricing.

### Pricing and Future Outlook

OpenAI offers competitive pricing for these advanced models. The o3 model is priced at $10 per million input tokens (equivalent to approximately 750,000 words) and $40 per million output tokens. The o4-mini model matches the pricing of its predecessor, o3-mini, at $1.10 per million input tokens and $4.40 for output tokens.

Looking ahead, OpenAI plans to introduce o3-pro, a premium variant that leverages heightened computational resources for generating responses, exclusively accessible to ChatGPT Pro subscribers.

### Final Standalone Models

OpenAI’s CEO Altman has hinted that o3 and o4-mini might be the final standalone AI reasoning models released before the anticipated GPT-5 launch. GPT-5 is expected to unify existing models like GPT-4.1 with the new reasoning models.

### About the Author

_The article was reported by Maxwell Zeff, a senior reporter at TechCrunch with an expertise in AI and emerging technologies. Zeff has previously worked with Gizmodo, Bloomberg, and MSNBC, covering critical developments in AI and technology i.e., the Silicon Valley Bank crisis. Zeff is based in San Francisco, exploring the Bay Area’s nature and culinary offerings in his free time._