DALL-E 2 AI Image Editing Model

$$ · 1.6¢ Deprecated

OpenAI's legacy image generation model supporting generations, edits with masks (inpainting), and variations

1024 x 1024
Max Resolution
10
Max Images per Request
Input / Output
Text Image Image
Deprecated

Details

Model ID
dall-e-2
Creator
OpenAI
Family
dall-e
Sunset
May 12, 2026
Tags
image-generation text-to-image image-editing inpainting
// Get Started

Ready to integrate?

Access dall-e-2 via our unified API.

Create Account

Providers & Pricing (2)

DALL-E 2 is available from 2 providers, with per-image pricing starting at $0.016 through Replicate.

Replicate
replicate/dall-e-2
Provider Model ID: openai/dall-e-2
$0.020 /image
OpenAI
openai/dall-e-2
Provider Model ID: dall-e-2

Output

Image 1024x1024
$0.020 per image
Image 256x256
$0.016 per image
Image 512x512
$0.018 per image
Pricing Notes (3)
  • Deprecated model - will stop being supported on May 12, 2026
  • Pricing is per image, varying by size
  • Supports generations, edits with masks (inpainting), and variations

DALL-E 2 API OpenAI-compatible

Lumenfall provides an OpenAI-compatible API for professional image generation, deep inpainting using masks, and the creation of visual variations via DALL-E 2.

Base URL
https://api.lumenfall.ai/openai/v1
Model
dall-e-2

Text to Image Generate

Create images from text descriptions

curl -X POST \
  https://api.lumenfall.ai/openai/v1/images/generations \
  -H "Authorization: Bearer $LUMENFALL_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "dall-e-2",
    "prompt": "A serene mountain landscape at sunset",
    "size": "1024x1024"
  }'
# Response:
# { "created": 1234567890, "data": [{ "url": "https://...", "revised_prompt": "..." }] }

Image Editing Edit

Transform existing images with text instructions

curl -X POST \
  https://api.lumenfall.ai/openai/v1/images/edits \
  -H "Authorization: Bearer $LUMENFALL_API_KEY" \
  -F "model=dall-e-2" \
  -F "[email protected]" \
  -F "prompt=Add a starry night sky to this image" \
  -F "size=1024x1024"
# Response:
# { "created": 1234567890, "data": [{ "url": "https://...", "revised_prompt": "..." }] }

DALL-E 2 FAQ

How much does DALL-E 2 cost?

DALL-E 2 starts at $0.016 per image through Lumenfall. Pricing varies by provider. Lumenfall does not add any markup to provider pricing.

How do I use DALL-E 2 via API?

You can use DALL-E 2 through Lumenfall's OpenAI-compatible API. Send requests to the unified endpoint with model ID "dall-e-2". Code examples are available in Python, JavaScript, and cURL.

Which providers offer DALL-E 2?

DALL-E 2 is available through Replicate and OpenAI on Lumenfall. Lumenfall automatically routes requests to the best available provider.

What is the maximum resolution for DALL-E 2?

DALL-E 2 supports images up to 1024x1024 resolution.

Overview

DALL-E 2 is a legacy text-to-image diffusion model developed by OpenAI that generates images from natural language descriptions. While succeeded by newer iterations, it remains a stable benchmark for image synthesis, offering a distinct feature set that includes image-to-image variations and mask-based inpainting. It is particularly known for its ability to combine disparate concepts and objects in a coherent, albeit often stylized, visual manner.

Strengths

  • Image Inpainting: The model excels at modifying existing images through masking, allowing users to replace specific elements or extend backgrounds while maintaining the original image’s context and lighting.
  • Concept Blending: It demonstrates a strong capability for semantic synthesis, such as placing a 3D-rendered character in a real-world setting or applying specific artistic styles (e.g., “in the style of Van Gogh”) to original subjects.
  • Compositional Understanding: DALL-E 2 handles spatial relationships and object attributes with reasonable accuracy, ensuring that adjectives are generally applied to the correct nouns within a prompt.
  • Variation Generation: It can ingest an existing image and output multiple visual permutations that retain the original’s core theme and color palette without being exact copies.

Limitations

  • Low Resolution: Native output is limited to 1024x1024 pixels, which often lacks the fine-grained texture and sharp detail found in more modern models like DALL-E 3 or Midjourney.
  • Text Rendering: The model struggle significantly with rendering legible text; characters often appear as nonsensical glyphs or blurred artifacts.
  • Photorealism Constraints: Compared to newer latent diffusion models, DALL-E 2 often produces images with a “plastic” or overly smooth aesthetic, struggling with complex human anatomy like hands or eyes.

Technical Background

DALL-E 2 is built on a CLIP-guided diffusion architecture, specifically a process OpenAI refers to as “unCLIP.” It uses the CLIP (Contrastive Language-Image Pre-training) latent space to translate text embeddings into image embeddings, which a decoder then converts into a visual representation. This approach prioritizes the relationship between visual concepts and their linguistic descriptions over raw pixel-mapping.

Best For

DALL-E 2 is best suited for rapid prototyping, creating stylized illustrations, and performing basic image editing tasks like inpainting or outpainting where high-fidelity photorealism isn’t the primary requirement. It is a cost-effective choice for developers who need consistent, programmatic image variations.

This model is available for testing and integration through Lumenfall’s unified API and interactive playground, allowing you to compare its outputs directly against more recent generative models.

Try DALL-E 2 in Playground

Generate images with custom prompts — no API key needed.

Open Playground