Alibaba's Wan 2.7 image generation and editing model for text-to-image, reference-guided generation, and instruction-based image edits
Details
wan-2.7
Starting from
Prices shown are in USD
Full pricing detailsProvider Performance
Fastest generation through alibaba at 6,235ms median latency with 84.0% success rate.
Aggregated from real API requests over the last 30 days.
Generation Time
Success Rate
Time to First Byte
Provider Rankings
| # | Provider | p50 Gen Time | p95 Gen Time | Success Rate | TTFB (p50) |
|---|---|---|---|---|---|
| 1 | alibaba | 6,235ms | 27,056ms | 84.0% | 5,702ms |
Providers & Pricing (1)
Wan 2.7 is available exclusively through Alibaba Cloud, starting at $0.03/image.
All modes
alibaba/wan-2.7-image
wan2.7-image API OpenAI-compatible
Access Wan 2.7 through Lumenfall’s OpenAI-compatible API to programmatically generate high-fidelity images and execute complex reference-guided image edits.
https://api.lumenfall.ai/openai/v1
wan-2.7
Code Examples
Text to Image
/v1/images/generationscurl -X POST \
https://api.lumenfall.ai/openai/v1/images/generations \
-H "Authorization: Bearer $LUMENFALL_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "wan-2.7",
"prompt": "",
"size": "1024x1024"
}'
# Response:
# { "created": 1234567890, "data": [{ "url": "https://...", "revised_prompt": "..." }] }
import OpenAI from 'openai';
const client = new OpenAI({
apiKey: 'YOUR_API_KEY',
baseURL: 'https://api.lumenfall.ai/openai/v1'
});
const response = await client.images.generate({
model: 'wan-2.7',
prompt: '',
size: '1024x1024'
});
// { created: 1234567890, data: [{ url: "https://...", revised_prompt: "..." }] }
console.log(response.data[0].url);
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://api.lumenfall.ai/openai/v1"
)
response = client.images.generate(
model="wan-2.7",
prompt="",
size="1024x1024"
)
# { created: 1234567890, data: [{ url: "https://...", revised_prompt: "..." }] }
print(response.data[0].url)
Image Edit
/v1/images/editsParameter Reference
Core Parameters
| Parameter | Type | Description | Modes |
|---|---|---|---|
prompt
|
string | Required. Text prompt for image generation |
T2I
Edit
|
Size & Layout
| Parameter | Type | Description | Modes |
|---|---|---|---|
size
|
string |
Image dimensions as WxH pixels (e.g. "1024x1024") or aspect ratio (e.g. "16:9")
WxH determines both shape and scale (aspect_ratio and resolution are ignored when size is provided). W:H format is equivalent to aspect_ratio.
|
T2I
Edit
|
aspect_ratio
|
string |
Aspect ratio of the output image (e.g. "16:9", "1:1")
Controls shape independently of scale. Use with resolution to control both. If size is also provided, size takes precedence. Any ratio is accepted and mapped to the nearest supported value.
|
T2I
Edit
|
resolution
|
string |
Output resolution tier (e.g. "1K", "4K")
Controls scale independently of shape. Higher tiers produce larger images and cost more. If size is also provided, size takes precedence for scale. Any tier is accepted and mapped to the nearest supported value.
|
T2I
Edit
|
size
Exact pixel dimensions
"1920x1080"
aspect_ratio
Shape only, default scale
"16:9"
resolution
Scale tier, preserves shape
"1K"
Priority when combined
size is most specific and always wins. aspect_ratio and resolution control shape and scale independently.
How matching works
7:1 on a model with
4:1 and 8:1,
you get 8:1.
0.5K 1K 2K 4K)
or megapixel tiers (0.25 1).
If the exact tier isn't available, you get the nearest one.
Media Inputs
| Parameter | Type | Description | Modes |
|---|---|---|---|
image
|
file |
Required.
Input image(s) to edit
Supports PNG, JPEG, WebP.
|
T2I
Edit
|
Output & Format
| Parameter | Type | Description | Modes |
|---|---|---|---|
response_format
|
string |
How to return the image
url
b64_json
Default:
"url" |
T2I
Edit
|
output_format
|
string |
Output image format
png
jpeg
gif
webp
avif
Gateway converts to requested format if provider doesn't support it natively.
|
T2I
Edit
|
output_compression
|
integer | Compression level for lossy formats (JPEG, WebP, AVIF) |
T2I
Edit
|
n
|
integer |
Number of images to generate
Default:
1Gateway generates multiple images in parallel even if provider only supports 1.
|
T2I
Edit
|
Parameter Normalization
How we handle parameters across different providers
Not every provider speaks the same language. When you send a parameter, we handle it in one of four ways depending on what the model supports:
| Behavior | What happens | Example |
|---|---|---|
passthrough |
Sent as-is to the provider | style, quality |
renamed |
Same value, mapped to the field name the provider expects | prompt |
converted |
Transformed to the provider's native format | size |
emulated |
Works even if the provider has no concept of it | n, response_format |
Parameters we don't recognize pass straight through to the upstream API, so provider-specific options still work.
Gallery
View all 4 imagesWan 2.7 FAQ
How much does Wan 2.7 cost?
Wan 2.7 starts at $0.03 per image through Lumenfall. Pricing varies by provider. Lumenfall does not add any markup to provider pricing.
How do I use Wan 2.7 via API?
You can use Wan 2.7 through Lumenfall's OpenAI-compatible API. Send requests to the unified endpoint with model ID "wan-2.7". Code examples are available in Python, JavaScript, and cURL.
Which providers offer Wan 2.7?
Wan 2.7 is available through Alibaba Cloud on Lumenfall. Lumenfall automatically routes requests to the best available provider.
What is the maximum resolution for Wan 2.7?
Wan 2.7 supports images up to 2048x2048 resolution.
Overview
Wan 2.4 is an image generation and editing model developed by Alibaba. It is designed to bridge the gap between pure text-to-image synthesis and precise, instruction-based image manipulation. The model is distinctive for its natively integrated support for multiple input modalities, allowing users to generate high-fidelity visuals from text prompts, use reference images to guide the aesthetic, or perform complex edits on existing images using natural language instructions.
Strengths
- Instruction-Based Editing: The model excels at following precise linguistic instructions to modify existing images, such as adding objects, changing backgrounds, or altering specific attributes while maintaining the integrity of the original composition.
- Reference-Guided Synthesis: Wan 2.4 demonstrates high fidelity when using external images as visual anchors, ensuring that the generated output retains stylistic or structural consistency with the provided reference material.
- Semantic Alignment: It exhibits strong prompt adherence, accurately translating complex or multi-part text descriptions into coherent visual scenes with minimal artifacting in the primary subjects.
- Multi-Modal Versatility: Unlike models restricted to a single input type, Wan 2.4 handles text-to-image, image-to-image, and reference-guided generation within a single framework, streamlining workflows that require iterative refinement.
Limitations
- Sequential Editing Sensitivity: When performing multiple rounds of instruction-based edits, the model may occasionally introduce “drift,” where the original image’s fine details gradually lose consistency over repeated transformations.
- Contextual Complexity: While strong at following instructions, the model can struggle with highly technical or spatial layouts that involve more than four or five distinct interacting objects in a single frame.
Technical Background
Wan 2.4 belongs to the Wan family of generative models, utilizing a diffusion-based architecture optimized for multi-modal inputs. Alibaba implemented a unified latent space approach that treats text prompts and reference images as collaborative tokens, allowing the model to weight visual cues and linguistic instructions simultaneously. The training methodology focused on high-density datasets involving paired image-instruction sets to improve the model’s “intent recognition” during the editing process.
Best For
This model is ideal for creative professionals requiring iterative design workflows, such as rapid prototyping of marketing assets where a base image must be tweaked for different campaigns. It is also well-suited for developers building applications that require dynamic user-driven image modifications or style transfers.
Wan 2.4 is available through Lumenfall’s unified API and playground, providing a streamlined environment for testing text-to-image prompts and complex image-edit instructions in a single interface.
Try Wan 2.7 in Playground
Generate images with custom prompts — no API key needed.