gpt-image-1.5

Model Description

gpt-image-1.5-rev (0.1 per time)

gpt-image-1.5

OpenAI has officially launched a new version of ChatGPT Images, powered by its latest flagship image generation model, GPT Image 1.5. This new model is designed to be significantly more efficient, generating images up to four times faster than its predecessor. It is currently rolling out to all ChatGPT users, with a dedicated “Images” space accessible via the sidebar on both the web and mobile app. For developers, the model is available in the API as GPT Image 1.5, offering a 20% reduction in cost for both image inputs and outputs compared to the previous version.

A primary focus of this update is the ability to perform precise edits while maintaining the integrity of the original image. When users request changes to an uploaded photo, the model more reliably adheres to their intent, modifying only the specified areas while keeping elements like lighting, composition, and a person’s appearance consistent. This capability allows for practical applications such as virtual clothing or hairstyle try-ons and conceptual transformations that retain the essence of the source material. The model excels at various editing tasks, including adding, subtracting, blending, and transposing elements, effectively acting as a portable creative studio.

Technical improvements in instruction following and text rendering further distinguish GPT Image 1.5. The model can now handle intricate compositions, such as generating specific 6×6 grids of diverse objects, with much higher accuracy. Text rendering has also seen a significant leap, enabling the model to produce dense and small text found in infographics, diagrams, or code snippets with greater clarity. Additionally, the update enhances overall image quality, particularly in rendering natural-looking scenes and many small faces within a crowd, which were previously challenging for generative models.

To make the creative process more accessible, OpenAI has introduced a new Images feature within ChatGPT that includes preset filters and prompt suggestions. While the model shows marked improvements in scientific accuracy and visual vividness, OpenAI notes that it is not yet perfect; limitations remain in areas such as multilingual support and specific complex styling. Despite these hurdles, the model is already being utilized by enterprises for brand-consistent marketing graphics and e-commerce catalogs, where maintaining visual identity across multiple iterations is essential.

🔔How to Use

graph LR A("Purchase Now") --> B["Start Chat on Homepage"] A --> D["Read API Documentation"] B --> C["Register / Login"] C --> E["Enter Key"] D --> F["Enter Endpoint & Key"] E --> G("Start Using") F --> G style A fill:#f9f9f9,stroke:#333,stroke-width:1px style B fill:#f9f9f9,stroke:#333,stroke-width:1px style C fill:#f9f9f9,stroke:#333,stroke-width:1px style D fill:#f9f9f9,stroke:#333,stroke-width:1px style E fill:#f9f9f9,stroke:#333,stroke-width:1px style F fill:#f9f9f9,stroke:#333,stroke-width:1px style G fill:#f9f9f9,stroke:#333,stroke-width:1px

Purchase Now

Start Chat on Homepage

Register / Login

Enter Key

Read API Documentation

Enter Endpoint & Key

Start Using

Description Ends

Recommend Models

kimi-k2.5

Kimi K2.5 is a native multimodal model that significantly advances visual understanding and coding capabilities while introducing a revolutionary multi-agent swarm system for tackling complex, large-scale tasks.

DeepSeek-R1-all

Performance on par with OpenAI-o1, Fully open-source model & technical report, Code and models are released under the MIT License: Distill & commercialize freely.

claude-opus-4-1-20250805

Opus 4.1 advances our state-of-the-art coding performance to 74.5% on SWE-bench Verified. It also improves Claude’s in-depth research and data analysis skills, especially around detail tracking and agentic search.