Introduction to OpenAI’s Image-2 Models
OpenAI has initiated blind tests for two new image generation models, codenamed “Chestnut” and “Hazelnut,” as of December 9, 2025. This accelerated launch aims to counter Google’s recent advancements with Gemini 3, which has shown superior performance in various benchmarks. The models are expected to be marketed as Image-2 and Image-2-mini, addressing quality issues like color aberrations and text rendering that affected the previous GPT-Image-1 model.
The blind testing phase, typically occurring shortly before a major release, suggests that an official launch may coincide with the anticipated GPT-5.2 language model release. This development is significant in the competitive landscape of AI, where advancements in visual generation quality are crucial for market leadership.
Competitive Context: Responding to Google’s Gemini 3
OpenAI’s blind testing of its new image generation models, codenamed “Chestnut” and “Hazelnut,” reflects a strategic response to Google’s Gemini 3 Pro. Launched in late 2025, Gemini 3 Pro has showcased notable performance advantages, outperforming OpenAI’s GPT-5 Pro in critical areas such as complex reasoning and human preference tests. This competitive landscape has prompted OpenAI to initiate what has been described as an internal “code red” directive, underscoring the urgency to enhance its visual generation capabilities.
The importance of visual generation quality cannot be overstated in the current AI market, where differentiation is increasingly driven by performance metrics. OpenAI’s efforts to address past limitations, such as color accuracy and text rendering, are crucial for maintaining its position in an evolving landscape. As the company prepares for the imminent public release of these models, the focus remains on closing the performance gap highlighted by Gemini 3 Pro. This context sets the stage for a significant shift in the competitive dynamics of AI image generation.
Our Take: We’re watching how OpenAI’s advancements will influence the broader AI market, particularly in response to Google’s growing capabilities. The emphasis on quality in visual generation is vital for sustaining leadership in this increasingly competitive field.
Anatomy of Chestnut and Hazelnut (Image-2 and Image-2-mini)
OpenAI’s new image generation models, codenamed “Chestnut” (Image-2) and “Hazelnut” (Image-2-mini), introduce significant architectural and training improvements over the previous GPT-Image-1. These enhancements aim to address longstanding quality issues, particularly in color accuracy and text rendering, which have been critical pain points for users.
Key Differences
- Chestnut (Image-2): This flagship model focuses on high-resolution outputs and improved detail. It leverages advanced training techniques to enhance visual fidelity and reduce artifacts.
- Hazelnut (Image-2-mini): Designed for efficiency, this model offers quicker generation times while maintaining a decent level of quality, making it suitable for applications where speed is essential.
According to OpenAI, both models are equipped with targeted quality enhancements that allow for more nuanced image generation, including better handling of complex scenes and improved context awareness. These features are expected to set a new standard in the AI image generation landscape, positioning OpenAI to better compete with Google’s Gemini 3 Pro.
The blind testing phase, which began on December 9, 2025, indicates an imminent release, likely aligned with the upcoming GPT-5.2 launch, further intensifying the competition in the AI visual generation space.
Our Take: The introduction of Chestnut and Hazelnut represents a strategic move by OpenAI to reclaim leadership in the AI image generation market. We’re watching how these models perform in real-world applications and whether they can effectively meet user demands for quality and speed.
Key Capability Improvements in Image-2
OpenAI’s upcoming image generation models, Chestnut (Image-2) and Hazelnut (Image-2-mini), showcase notable improvements over the previous GPT-Image-1. These advancements are particularly evident in three key areas that enhance the user experience and output quality.
Enhanced Facial Realism and Portraits
The new models significantly improve facial realism, allowing for more lifelike representations in generated images. Users can expect celebrity-style portraits that capture intricate details and nuanced expressions, making them suitable for various applications, from marketing to entertainment.
Accurate Rendering of Technical Content
Image-2 models excel in rendering complex technical content, including accurate code and formulas. This improvement is crucial for professionals in fields requiring precise visual representations, such as engineering and scientific research.
Resolved Color Bias and Improved Image Fidelity
OpenAI claims that both models have addressed color bias issues that affected previous iterations. The result is enhanced image fidelity, ensuring that colors are rendered accurately and consistently across different contexts.
Overall, these enhancements position OpenAI’s Chestnut and Hazelnut models as formidable contenders in the AI image generation landscape, particularly as they prepare for their upcoming launch. As we monitor these developments, it remains to be seen how they will impact user adoption and competitive dynamics in the AI space.
Blind Testing Methodology and Evaluation Metrics
OpenAI’s blind testing of its new image generation models, Chestnut (Image-2) and Hazelnut (Image-2-mini), was conducted on the Design Arena and LM Arena platforms, starting December 9, 2025. These platforms are known for their rigorous evaluation of AI capabilities, allowing for a controlled comparison of model outputs.
The blind test setup involved generating images without revealing the model identity to the evaluators, ensuring unbiased assessments of image quality. Evaluation criteria focused on several key metrics, including color accuracy, detail resolution, and text rendering, addressing the shortcomings noted in the previous GPT-Image-1 model.
Initial results indicate significant scoring improvements over GPT-Image-1 samples. According to OpenAI, early evaluations suggest that Chestnut and Hazelnut outperform their predecessor in multiple aspects, with improvements noted in visual fidelity and coherence. This rigorous testing phase is critical as it precedes a potential public release, aligning with OpenAI’s strategy to enhance its competitive position in the rapidly evolving AI landscape.
Practical Applications and Use Cases
OpenAI’s new image generation models, Chestnut and Hazelnut, are poised to enhance various practical applications across industries. These models are particularly beneficial for content creation and marketing imagery, enabling businesses to generate high-quality visuals that are tailored to specific campaigns.
In technical documentation, Chestnut and Hazelnut can produce precise educational diagrams that clarify complex concepts, making learning more accessible. Furthermore, designers can leverage these models to create prototypes and concept artwork, streamlining the design process and fostering innovation.
The capabilities of these models extend beyond mere aesthetics; they are designed to improve workflow efficiency and reduce the time spent on visual content creation. As the demand for high-quality visuals continues to grow, the integration of advanced AI tools like Chestnut and Hazelnut is becoming increasingly essential for professionals across various fields.
Comparing Image-2 with Midjourney, DALL-E 3, Flux, and Canva
OpenAI’s Image-2 models, Chestnut and Hazelnut, are positioned against established players like Midjourney, DALL-E 3, Flux, and Canva. Key differentiators include:
- Quality: Image-2 aims to resolve prior issues in color accuracy and text rendering, outperforming DALL-E 3 in specific benchmarks.
- Speed: Midjourney and Flux are noted for their rapid generation times, while Image-2 is still being evaluated.
- Usability: Canva offers user-friendly interfaces, whereas Image-2 targets advanced users needing high-quality outputs.
Pros and Cons
- Image-2: High quality but may lack speed.
- Midjourney: Fast but can be inconsistent.
- DALL-E 3: Good quality but limited in customization.
- Flux: Fast and versatile.
- Canva: Accessible but less powerful for complex tasks.
When to choose Image-2? Opt for it when quality is paramount and advanced features are needed.
Launch Timing and Market Implications
OpenAI is reportedly preparing for the public launch of its new image generation models, Chestnut and Hazelnut, potentially coinciding with the release of the GPT-5.2 language model. This timing is crucial, as it allows OpenAI to enhance its competitive stance against Google’s Gemini 3, which has recently set new benchmarks in AI performance.
The blind testing phase for these models began on December 9, 2025, indicating a launch within weeks. According to industry insiders, the strategic importance of this release lies in its potential to address previous quality issues and to solidify OpenAI’s position in a rapidly evolving market. As the competition intensifies, the ability to deliver superior visual generation capabilities will be a key differentiator for OpenAI.
Conclusion and Future Outlook
OpenAI’s upcoming image generation models, Chestnut and Hazelnut, are poised to significantly enhance visual generation quality. These advancements address critical issues from the previous generation, particularly in color accuracy and text rendering. In the short term, users can expect improved image outputs, while developers will benefit from enhanced capabilities for integration into applications. Looking ahead, the models are set to influence the broader AI landscape, particularly as competition with Google’s Gemini 3 intensifies. Our take is that these developments will reshape user expectations and set new standards in AI-generated imagery.


