HappyHorse 1.0 Coming SoonLearn More

Wan 2.5 API

Turn short prompts or reference images into ready-to-post videos with audio output using Wan 2.5 API on Evolink AI.

Price: $0.0075 - 0.025(~ 0.51 - 1.7 credits) per second of video

Highest stability with guaranteed 99.9% uptime. Recommended for production environments.

Use the same API endpoint for all versions. Only the model parameter differs.

0 (suggested: 2,000)

480p supports 16:9, 9:16, 1:1; 720p/1080p add 4:3 and 3:4

Click Generate to see preview

History

Max 20 items

0 running · 0 completed

Your generation history will appear here

Wan 2.5 API for AI video with audio output

Generate short HD videos (5 or 10 seconds) with audio output and social-ready framing from simple text or image inputs, all via a clean Evolink AI API.

Social media showcase of AI video generation product feature 1

Pricing

WAN 2.5 Text to Video
Video Generation
Duration:5s
Quality:480p
Price:
$0.188/ video
(12.75 Credits)
WAN 2.5 Text to Video
Video Generation
Duration:5s
Quality:720p
Price:
$0.375/ video
(25.5 Credits)
WAN 2.5 Text to Video
Video Generation
Duration:5s
Quality:1080p
Price:
$0.626/ video
(42.585 Credits)
WAN 2.5 Text to Video
Video Generation
Duration:10s
Quality:480p
Price:
$0.375/ video
(25.5 Credits)
WAN 2.5 Text to Video
Video Generation
Duration:10s
Quality:720p
Price:
$0.750/ video
(51 Credits)
WAN 2.5 Text to Video
Video Generation
Duration:10s
Quality:1080p
Price:
$1.252/ video
(85.17 Credits)

If it's down, we automatically use the next cheapest available—ensuring 99.9% uptime at the best possible price.

What is Wan 2.5 API on Evolink AI?

Text-to-video at the cheapest Wan tier

Wan 2.5 API is the daily workhorse tier of the Wan video lineup, designed to keep per-second cost low while generating audio alongside video per current route docs. Send a short text prompt and receive a clip you can drop straight into a content calendar — the same per-second rate applies whether you generate one clip or batch hundreds, which is why teams reach for Wan 2.5 when they want sustainable daily volume rather than premium one-off shoots.

Vertical feed showcase of AI video generation product feature 2

Image-to-video for product stories

With Wan 2.5 API you can upload a single key visual, such as a product shot or character design, and turn it into a short, dynamic clip that still feels consistent with the original image. This works well for turning static catalog images into scroll-stopping ads, motion posters, or story snippets where the camera moves, the light changes, and audio output accompanies the visuals in a few seconds.

Product motion showcase of AI video generation product feature 3

Built for social-first creators

Wan 2.5 API via Evolink AI is designed around social media creators and indie SaaS builders who care about speed, volume, and consistency more than academic benchmarks. Per current route docs it supports 5 or 10 second clips, vertical and square formats, and audio output alongside the visuals, so your videos can drop straight into content calendars, UGC templates, or automated posting systems without hand-fixing every render.

Creator workflow showcase of AI video generation product feature 4

Why choose Wan 2.5 API via Evolink AI?

Wan 2.5 API combines Alibaba’s video model with Evolink AI’s simple routing so you focus on ideas, not cloud configs or complex infrastructure.

Video with audio output on current routes

Most AI video tools still make you juggle separate models for visuals and sound, which introduces friction and file management overhead. Wan 2.5 API generates audio alongside video per current route docs, so what you get already feels closer to a finished short-form clip. For creators and SaaS products that live or die by speed, that workflow removes a lot of invisible overhead.

Simple Wan 2.5 integration

Wan 2.5 itself is an Alibaba Wan model, but Evolink AI wraps it into a straightforward Wan 2.5 API so you do not need to deal with region settings, separate console projects, or complex billing dashboards. You call a single endpoint with clear parameters and Evolink AI handles routing to the underlying Wan 2.5 model, making it much easier for developers and creators to plug video generation into their products or content workflows.

A workhorse model for daily content volume

Wan 2.5 is the model teams reach for when they need to ship video every day rather than once per campaign. Because the per-second cost stays predictable across batches, you can budget for daily UGC pipelines, scheduled social posts, and SaaS features that generate clips on demand without each call turning into a separate finance conversation. It is less about chasing the absolute best frame and more about keeping creative output flowing at a sustainable unit cost.

How to use Wan 2.5 API

Connect Wan 2.5 API through Evolink AI and move from prompt to published video in a few simple steps.

1

Connect your Evolink AI account

Sign up or log into Evolink AI, create an API key, and enable Wan 2.5 API access so your app can securely call the video generation endpoints without touching Alibaba Cloud directly.

2

Send prompts, images, and basic settings

Choose text-to-video or image-to-video, write a clear prompt, upload an optional reference image, set duration and aspect ratio, then send a simple JSON request to the Wan 2.5 API route.

3

Receive, review, and publish your clips

Fetch the generated Wan 2.5 video URL, preview audio and visuals, then plug it into your editor, scheduler, or SaaS interface for immediate download, posting, or further automation.

Key Wan 2.5 API features

Wan 2.5 API on Evolink AI focuses on real-world social and marketing use cases rather than lab demos, so every feature maps to a clear creator benefit.

Audio output

Audio output alongside video on current routes

Wan 2.5 API renders audio alongside the visuals per current route docs, so your team can treat sound as a default part of the workhorse pipeline rather than a separate post-production step. For social and UGC volume work where most clips never get a manual sound pass, having audio output available keeps unit cost low and avoids drop-offs from silent clips that nobody finishes watching.

Short-form

Short HD clips for social feeds

The model is tuned for 5 or 10 second HD videos so you hit the sweet spot for TikTok, Reels, and ad placements without wasting budget or time on overly long renders that nobody watches.

Flexible input

Text or image input — no reference video needed

Wan 2.5 API focuses on the two inputs most daily content workloads actually use: a short text prompt or a single reference image. There is intentionally no reference-video mode at this tier — that role belongs to Wan 2.6 reference video — so the integration stays simple, the parameter surface stays small, and the per-second cost stays predictable for the high-volume social and UGC use cases Wan 2.5 was tuned for.

Global-ready

Multi-language prompt and audio support

Wan 2.5 API is comfortable with Chinese and English prompts, which is especially useful when your audience spans multiple regions and languages online.

Cinematic feel

Consistent motion and control

The model offers smoother motion and better camera dynamics than older Wan versions, so videos feel more cinematic and less like janky demos, even when you move fast on campaigns.

Budget control

Per-second budget control for daily SaaS workloads

Because Wan 2.5 API on Evolink AI runs at a single per-second rate, you can wire it into cron jobs, no-code tools, or SaaS backends and predict the daily cost of generating N clips before you even ship the feature. That predictability is what makes Wan 2.5 a viable workhorse model for content SaaS, UGC platforms, and scheduled social feeds where the finance side cares about unit cost more than peak quality.

Explore the Wan API family

Wan 2.5 is the daily workhorse tier in the Wan family, with audio output supported on current routes and a predictable per-second cost. For cinematic multi-shot output or reference video, move up to Wan 2.6. For text-to-image and image-to-image, use Wan Image on the same Wan account.

Wan 2.5 API FAQ

Everything you need to know about the product and billing.

Wan 2.5 is an AI video model developed by Alibaba’s Tongyi Wanxiang team that turns text prompts or single images into short HD videos, with audio output supported on current routes. Wan 2.5 API via Evolink AI is an integration layer that exposes this model through one user-friendly API, so you can call Wan 2.5 from your apps, SaaS products, or creator tools without touching Alibaba Cloud directly. It is designed for developers, marketers, and social media creators who want reliable video generation as a service rather than managing low-level infrastructure.
Wan 2.5 API is best at short, story-like clips that fit social platforms and performance ads. You can use it to produce product explainers, scroll-stopping intros, motion posters, and quick story beats that pair visuals with sound. Many teams use Wan 2.5 as a way to test hooks and concepts before investing in full productions, generating multiple variants of the same idea and letting real engagement data decide which direction to scale.
Wan 2.5 API delivers a strong balance of video quality, cost, and speed for 5 or 10 second videos, while newer models such as Wan 2.6 push further into multi-shot storytelling and advanced narrative control. For many day-to-day social and UGC tasks, Wan 2.5 already provides enough realism and motion to feel native in feeds, at a friendlier price point than heavy cinematic models. That makes it a good fit as a workhorse model in your stack, especially when combined with higher-end options for premium shots.
Wan 2.5 API supports both text-to-video and image-to-video modes, which you can choose through a parameter in the Evolink AI endpoint. In text mode, you provide a prompt describing scene, style, and pacing, and the model generates a fresh clip from scratch, with audio output supported on current routes. In image mode, you upload a still frame, key visual, or character and let the video animate around it, preserving identity while adding camera moves and transitions.
Per current route docs, Wan 2.5 API clips use documented duration choices of 5 seconds or 10 seconds, which is ideal for hooks, transitions, and bite-sized stories on social. Resolution options commonly span 480p, 720p, and 1080p, depending on your chosen settings and plan. This keeps render times and costs manageable while still delivering visuals that look clean on modern phone screens and ad placements in vertical or square formats.
Wan 2.5 API is typically billed on a per-second basis through usage-based credits, with pricing aligned to roughly mid-market levels for HD video generation. This means shorter clips and lower resolutions use fewer credits, while longer or higher-resolution clips consume more, giving you control over budget and quality. The result is a pay-for-what-you-generate model that works well for experiments, campaigns, and SaaS workloads with changing demand.
Wan 2.5 itself is a paid AI video generator from Alibaba's Wan lineup, not a free model, but you can usually try it without committing to a long contract by signing up for Evolink AI and using introductory credits to generate your first Wan 2.5 image-to-video or text-to-video clips. For teams looking for the cheapest Wan 2.5 access, the practical path is to compare the per-second cost of Wan 2.5 on Evolink against the official DashScope rate and choose the route that fits your volume. Use the Wan 2.5 playground or preview to validate quality before scaling up, and only top up credits once you have confirmed Wan 2.5 fits your social, ad, or UGC workflow.
Alibaba open-sourced earlier Wan releases such as Wan 2.1, while Wan 2.5 is documented as an API-accessible model on Alibaba's DashScope and Model Studio. As of April 2026, we have not found an official Alibaba source confirming Wan 2.5 itself as open source, so for the most current status please check Alibaba's official announcements. To use Wan 2.5 today, you can call it via the Wan 2.5 API on Evolink AI without managing Alibaba Cloud infrastructure directly.
Wan 2.5 API via Evolink AI is designed with commercial use in mind, from performance ads to branded content and embedded SaaS features. As with any AI service, you should still review the latest terms from both Evolink AI and the underlying provider, check usage guidelines for your region and vertical, and put a simple internal review step in place before launching client campaigns or high-visibility assets.
You can call Wan 2.5 API from backend services, serverless functions, low-code tools, or even no-code automation platforms that support HTTP requests. A common pattern is to trigger video generation when new content is created, when a user submits a brief, or on a schedule for recurring campaigns. Evolink AI’s documentation includes example requests, response objects, and error patterns so your engineering time goes into product logic instead of debugging raw infrastructure.
POST
/v1/videos/generations

Create Video

Wan 2.5 Video Text to Video (wan2.5-text-to-video) model supports text-to-video generation mode.

Asynchronous processing mode, use the returned task ID to .

Generated video links are valid for 24 hours, please save them promptly.

Request Parameters

modelstringRequiredDefault: wan2.5-text-to-video

Video generation model name.

Examplewan2.5-text-to-video
promptstringRequired

Prompt describing what kind of video to generate.

Notes
  • Limited to 2000 tokens
ExampleA cat playing piano
aspect_ratiostringOptionalDefault: 16:9

Video aspect ratio.

ValueDescription
16:9Landscape video
9:16Portrait video
1:1Square video
4:3Standard video (720p/1080p only)
3:4Portrait standard (720p/1080p only)
Notes
  • 480p supports: 16:9, 9:16, 1:1
  • 720p/1080p supports all ratios
Example16:9
qualitystringOptionalDefault: 720p

Video quality.

ValueDescription
480pLower quality, lower price
720pStandard quality (default)
1080pHigh quality, higher price
Notes
  • Different quality levels support different aspect ratios
Example720p
durationintegerOptional

Duration of the generated video (seconds).

ValueDescription
55 seconds
1010 seconds
Notes
  • Pre-charged based on duration, actual charge based on generated video duration
Example5
prompt_extendbooleanOptionalDefault: true

Whether to enable intelligent prompt rewriting.

Notes
  • When enabled, a large language model will optimize the prompt
  • Effective for prompts that lack detail or are too simple
Exampletrue
callback_urlstringOptional

HTTPS callback address after task completion.

Notes
  • Triggered on completion, failure, or cancellation
  • Sent after billing confirmation
  • HTTPS only, no internal IPs
  • Max length: 2048 chars
  • Timeout: 10s, Max 3 retries
Examplehttps://your-domain.com/webhooks/video-task-completed

Request Example

{
  "model": "wan2.5-text-to-video",
  "prompt": "A cat playing piano",
  "aspect_ratio": "16:9",
  "quality": "720p",
  "duration": 5
}

Response Example

{
  "created": 1757169743,
  "id": "task-unified-1757169743-7cvnl5zw",
  "model": "wan2.5-text-to-video",
  "object": "video.generation.task",
  "progress": 0,
  "status": "pending",
  "task_info": {
    "can_cancel": true,
    "estimated_time": 30,
    "video_duration": 8
  },
  "type": "video",
  "usage": {
    "billing_rule": "per_call",
    "credits_reserved": 5,
    "user_group": "default"
  }
}
Wan 2.5 API: AI Video Generator with Audio Output | Evolink AI