Qwen Image Edit 2509

AI Image Editing Model

Image $$ · 3¢

Alibaba's Qwen image editing model for instruction-based image modifications and transformations

Example outputs coming soon

Supported Modes
Image Edit
Active

Details

Model ID
qwen-image-edit-2509
Also known as: qwen-image-edit-plus-2509
Creator
Family
qwen
Released
September 2025
Tags
image-generation image-editing
// Get Started

Ready to integrate?

Access qwen-image-edit-2509 via our unified API.

Create Account
Available at 1 provider

Starting from

$0.030 /image via fal.ai

Prices shown are in USD · Some prices estimated from per-megapixel or per-token pricing

Full pricing details

Providers & Pricing (1)

Qwen Image Edit 2509 is available exclusively through fal.ai, starting at $0.03/image.

fal.ai
fal/qwen-image-edit-2509
Provider Model ID: fal-ai/qwen-image-edit-2509
$0.030 /megapixel

qwen-image-edit-plus-2509 API OpenAI-compatible

Integrate Qwen Image Edit 2509 into workflows via Lumenfall's OpenAI-compatible API to perform text-to-image generation and complex image transformations using natural language instructions.

Base URL
https://api.lumenfall.ai/openai/v1
Model
qwen-image-edit-2509

Code Examples

Image Edit

/v1/images/edits
curl -X POST \
  https://api.lumenfall.ai/openai/v1/images/edits \
  -H "Authorization: Bearer $LUMENFALL_API_KEY" \
  -F "model=qwen-image-edit-2509" \
  -F "[email protected]" \
  -F "prompt=Add a starry night sky to this image" \
  -F "size=1024x1024"
# Response:
# { "created": 1234567890, "data": [{ "url": "https://...", "revised_prompt": "..." }] }

Parameter Reference

Required Supported Not available

Core Parameters

Parameter Type Description Modes
prompt string Required. Edit instruction for the image
Edit
negative_prompt string Negative prompt to guide generation away from undesired content
Edit
seed integer Random seed for reproducibility
Edit

Size & Layout

Parameter Type Description Modes
size string Image dimensions as WxH pixels (e.g. "1024x1024") or aspect ratio (e.g. "16:9")
WxH determines both shape and scale (aspect_ratio and resolution are ignored when size is provided). W:H format is equivalent to aspect_ratio.
Edit
aspect_ratio string Aspect ratio of the output image (e.g. "16:9", "1:1")
Controls shape independently of scale. Use with resolution to control both. If size is also provided, size takes precedence. Any ratio is accepted and mapped to the nearest supported value.
Edit
resolution string Output resolution tier (e.g. "1K", "4K")
Controls scale independently of shape. Higher tiers produce larger images and cost more. If size is also provided, size takes precedence for scale. Any tier is accepted and mapped to the nearest supported value.
Edit
size

Exact pixel dimensions

"1920x1080"
aspect_ratio

Shape only, default scale

"16:9"
resolution

Scale tier, preserves shape

"1K"

Priority when combined

size aspect_ratio + resolution aspect_ratio resolution

size is most specific and always wins. aspect_ratio and resolution control shape and scale independently.

How matching works

Shape matching – we pick the closest supported ratio. Ask for 7:1 on a model with 4:1 and 8:1, you get 8:1.
Scale matching – providers use different tier formats: K tiers (0.5K 1K 2K 4K) or megapixel tiers (0.25 1). If the exact tier isn't available, you get the nearest one.
Dimension clamping – if a model has pixel limits, we clamp dimensions to fit and keep the aspect ratio intact.

Media Inputs

Parameter Type Description Modes
image file Required. Input image(s) to edit
Supports PNG, JPEG, WebP.
Edit

Output & Format

Parameter Type Description Modes
response_format string How to return the image
url b64_json
Default: "url"
Edit
output_format string Output image format
png jpeg gif webp avif
Gateway converts to requested format if provider doesn't support it natively.
Edit
output_compression integer Compression level for lossy formats (JPEG, WebP, AVIF)
Edit
n integer Number of images to generate
Default: 1
Gateway generates multiple images in parallel even if provider only supports 1.
Edit

Additional Parameters

Parameter Type Description Modes
cfg_scale number Classifier-free guidance scale — higher values stick more closely to the prompt
Edit
acceleration fal string Acceleration level for image generation. Options: 'none', 'regular'. Higher acceleration increases speed. 'regular' balances speed and quality.
none regular
Edit
enable_safety_checker fal boolean If set to true, the safety checker will be enabled.
Edit
num_inference_steps fal integer The number of inference steps to perform.
Edit
sync_mode fal boolean If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
Edit

Parameter Normalization

How we handle parameters across different providers

Not every provider speaks the same language. When you send a parameter, we handle it in one of four ways depending on what the model supports:

Behavior What happens Example
passthrough Sent as-is to the provider style, quality
renamed Same value, mapped to the field name the provider expects prompt
converted Transformed to the provider's native format size
emulated Works even if the provider has no concept of it n, response_format

Parameters we don't recognize pass straight through to the upstream API, so provider-specific options still work.

Qwen Image Edit 2509 FAQ

How much does Qwen Image Edit 2509 cost?

Qwen Image Edit 2509 starts at $0.03 per image through Lumenfall. Pricing varies by provider. Lumenfall does not add any markup to provider pricing.

How do I use Qwen Image Edit 2509 via API?

You can use Qwen Image Edit 2509 through Lumenfall's OpenAI-compatible API. Send requests to the unified endpoint with model ID "qwen-image-edit-2509". Code examples are available in Python, JavaScript, and cURL.

Which providers offer Qwen Image Edit 2509?

Qwen Image Edit 2509 is available through fal.ai on Lumenfall. Lumenfall automatically routes requests to the best available provider.

Overview

Qwen Image Edit 2509 is a specialized vision-language model developed by Alibaba designed for instruction-based image manipulation. Unlike standard text-to-image generators, this model accepts both an initial image and a natural language prompt to perform targeted modifications and transformations. It is distinctive for its ability to interpret complex editing instructions while maintaining the structural integrity of the original source image.

Strengths

  • Instruction Adherence: The model accurately maps natural language verbs and nouns to visual changes, such as “change the color of the shirt” or “add a sunset to the background.”
  • Contextual Consistency: It excels at preserving the identity and spatial layout of primary subjects while altering specific attributes or environmental elements.
  • Zero-shot Composition: The model handles various editing tasks—including stylization, object insertion, and attribute modification—without requiring mask-based inputs or fine-tuning for specific styles.
  • Complex Transformation: Beyond simple filters, it can handle structural transformations such as changing a character’s pose or modifying the lighting conditions of a scene based on text descriptions.

Limitations

  • High-Detail Text Rendering: Like many diffusion-based or vision-language architectures, it may struggle with rendering precise, small-scale legible text within an edited image.
  • Large-Scale Compositional Overhauls: While it handles local edits and style transfers well, it may produce artifacts if the prompt asks for a complete reimagining that contradicts the fundamental geometry of the source image.
  • Anatomical Accuracy: In complex edits involving human figures, there is a risk of generating anatomical inconsistencies, particularly in hands or overlapping limbs.

Technical Background

Developed as part of the Qwen model family, Qwen Image Edit 2509 utilizes a vision-encoder paired with a generative backbone trained on large-scale paired datasets of images and their corresponding edit instructions. The architecture focuses on cross-modal alignment, ensuring that the text embeddings effectively guide the latent representation of the source image during the denoising or reconstruction process. This approach prioritizes semantic understanding of the “before” and “after” relationship described in the prompt.

Best For

Qwen Image Edit 2509 is best suited for workflows requiring rapid prototyping of visual concepts, such as changing product backgrounds, adjusting fashion photography attributes, or iterative character design. It is an excellent choice for developers building creative tools that require “natural language” control over existing visual assets rather than generating images from scratch.

This model is available through Lumenfall’s unified API and playground, allowing for easy integration into multi-model pipelines alongside text and vision-analysis models.

Try Qwen Image Edit 2509 in Playground

Generate images with custom prompts — no API key needed.

Open Playground