Home/Models/OpenAI/GPT Image 1 mini
O

GPT Image 1 mini

輸入:$6.4/M
輸出:$25.6/M
Cost-optimized version of GPT Image 1. It is a native Multimodal language model that accepts both text and image input and generates image output.
新
商用
概览
功能亮点
定价
API

Key features

  • Text→Image generation: converts natural-language prompts into images with strong instruction following.
  • Image editing / inpainting: accepts reference images and masks to perform targeted edits.
  • Cost-optimized (“mini”) design: a smaller footprint that OpenAI and observers describe as much cheaper per image than the large model (OpenAI/DevDay messaging and early reports say ~80% less expensive).
  • Flexible output controls: supports size, output format (JPEG/PNG/WEBP), compression and a quality knob (low/medium/high/auto in the cookbook).

Technical details (architecture & capabilities)

  • Model family & input/output: member of the gpt-image-1 family; accepts text prompts and image inputs (for edits) and returns generated image outputs. Quality/size parameters control resolution (typical max ~1536×1024 in this family—see docs for exact supported sizes).
  • Operational tradeoffs: engineered as a smaller footprint model—trades some top-end fidelity for throughput and cost improvements while preserving robust prompt-following and edit features.
  • Safety & metadata: follows OpenAI’s image safety guardrails and embeds C2PA metadata options for provenance when available.

Inputs & outputs — canonical usage supports:

  • Text prompt (string) to generate a new image.
  • Image + mask to perform targeted edits/inpainting.
  • Reference images to control style or composition.
    These are exposed via the Images API (model name gpt-image-1-mini).

Limitations

  • Lower peak fidelity: compared with the large gpt-image-1 model, mini may lose some micro-detail and top-end photorealism (expected tradeoff for cost).
  • Text rendering & tiny details: like many image models, it can struggle with small legible text, dense charts, or micro-fine textures; expect to post-process or use higher-capacity models for those needs.
  • Edit scope: image edit/inpainting features are available but suggest some editing limitations relative to interactive ChatGPT web tools—edits are effective for many tasks but may require iterative refinement.
  • Safety & policy constraints: outputs are subject to OpenAI moderation/safety guardrails (explicit content, copyrighted content restrictions, disallowed outputs). Developers can control moderation sensitivity via API parameters where offered.

Recommended use cases

  • High-volume content generation (marketing assets, thumbnails, rapid concept art) — where cost per image is primary.
  • Programmatic editing / templating — bulk inpainting or variant generation from a base asset.
  • Interactive applications with budget constraints — chat interfaces or integrated design tools where response speed and cost matter more than absolute top fidelity.
  • Prototyping & A/B image generation — generate many candidate images quickly and selectively upscale or re-run on larger models for finalists.
  • How to access gpt-image-1-mini API

Step 1: Sign Up for API Key

Log in to cometapi.com. If you are not our user yet, please register first. Sign into your CometAPI console. Get the access credential API key of the interface. Click “Add Token” at the API token in the personal center, get the token key: sk-xxxxx and submit.

img

Step 2: Send Requests to gpt-image-1-mini API

Select the “\**gpt-image-1-mini \**”endpoint to send the API request and set the request body. The request method and request body are obtained from our website API doc. Our website also provides Apifox test for your convenience. Replace <YOUR_API_KEY> with your actual CometAPI key from your account.

Insert your question or request into the content field—this is what the model will respond to . Process the API response to get the generated answer.

Step 3: Retrieve and Verify Results

Process the API response to get the generated answer. After processing, the API responds with the task status and output data.

GPT Image 1 mini 的功能

了解 GPT Image 1 mini 的核心能力,帮助提升性能与可用性,并改善整体体验。

GPT Image 1 mini 的定价

查看 GPT Image 1 mini 的竞争性定价,满足不同预算与使用需求,灵活方案确保随需求扩展。
Comet 价格 (USD / M Tokens)官方定价 (USD / M Tokens)折扣
輸入:$6.4/M
輸出:$25.6/M
輸入:$8/M
輸出:$32/M
-20%

GPT Image 1 mini 的示例代码与 API

gpt-image-1-mini is a cost-optimized, multimodal image model from OpenAI that accepts text and image inputs and produces image outputs. It is positioned as a smaller, cheaper sibling to OpenAI’s full GPT-Image-1 family — designed for high-throughput production use where cost and latency are important constraints. The model is intended for tasks such as text-to-image generation, image editing / inpainting, and workflows that incorporate reference imagery.
Python
JavaScript
Curl
import base64
import os
from openai import OpenAI
from PIL import Image
from io import BytesIO

# Get your CometAPI key from https://api.cometapi.com/console/token, and paste it here
COMETAPI_KEY = os.environ.get("COMETAPI_KEY") or "<YOUR_COMETAPI_KEY>"
client = OpenAI(api_key=COMETAPI_KEY, base_url="https://api.cometapi.com/v1")

# Output directory
OUTPUT_DIR = os.path.join(os.path.dirname(__file__), "..", "output")
os.makedirs(OUTPUT_DIR, exist_ok=True)

prompt = "A cute baby sea otter swimming in the ocean"

# Generate image using gpt-image-1-mini
result = client.images.generate(
    model="gpt-image-1-mini",
    prompt=prompt,
    size="1024x1024",
)

# Save the image
image_base64 = result.data[0].b64_json
image_bytes = base64.b64decode(image_base64)

output_path = os.path.join(OUTPUT_DIR, "output.png")
image = Image.open(BytesIO(image_bytes))
image.save(output_path, format="PNG")

print(f"Image saved to: {output_path}")

更多模型