each::sense is in private beta.
Eachlabs | AI Workflows for app builders

WAN-2.5

Wan 2.5 Preview is a model that generates short, cinematic videos from a single input image. It preserves the details of the original image while adding camera movements and atmosphere to bring the scene to life. This allows a still photo to be transformed into a film-like moving sequence. The “Preview” version is optimized for quick tests and concept exploration, making it ideal for prototyping and creative experimentation.

Avg Run Time: 385.000s

Model Slug: wan-2-5-preview-image-to-video

Playground

Input

Enter a URL or choose a file from your computer.

Enter a URL or choose a file from your computer.

Advanced Controls

Output

Example Result

Preview and download your result.

Unsupported conditions - pricing not available for this input format

API & SDK

Create a Prediction

Send a POST request to create a new prediction. This will return a prediction ID that you'll use to check the result. The request should include your model inputs and API key.

Get Prediction Result

Poll the prediction endpoint with the prediction ID until the result is ready. The API uses long-polling, so you'll need to repeatedly check until you receive a success status.

Readme

Table of Contents
Overview
Technical Specifications
Key Considerations
Tips & Tricks
Capabilities
What Can I Use It For?
Things to Be Aware Of
Limitations

Overview

Wan 2.5 Preview is an advanced image-to-video generative AI model designed to transform a single input image into a short, cinematic video sequence. Developed by the Wan research team, this model leverages state-of-the-art generative techniques to add dynamic camera movements, atmospheric effects, and subtle scene enhancements while preserving the core details and composition of the original image. The primary goal is to breathe life into still photos, making them appear as if they are part of a film or moving scene.

The “Preview” version of Wan 2.5 is specifically optimized for rapid prototyping and creative experimentation. It is engineered for speed and responsiveness, allowing users to quickly test concepts and iterate on visual ideas without the overhead of full production rendering. The underlying architecture combines advanced diffusion models with temporal consistency modules, ensuring smooth transitions and realistic motion effects. What sets Wan 2.5 Preview apart is its ability to maintain high fidelity to the input image while introducing cinematic elements, making it highly valuable for artists, designers, and creative professionals seeking to visualize motion from static imagery.

Technical Specifications

  • Architecture: Diffusion-based generative model with temporal consistency enhancements
  • Parameters: Not publicly disclosed; estimated to be in the hundreds of millions based on comparable models
  • Resolution: Supports input images up to 1024x1024 pixels; output video resolution typically matches or slightly exceeds input resolution
  • Input/Output formats: Accepts standard image formats (JPEG, PNG); outputs short video clips in MP4 or GIF format
  • Performance metrics: Average generation time per video ranges from 10 to 30 seconds depending on hardware; maintains high perceptual similarity scores between input image and generated frames

Key Considerations

  • The model excels at generating cinematic camera movements and atmospheric effects but may introduce minor artifacts if the input image is low quality or highly complex
  • For best results, use high-resolution, well-lit images with clear subject separation
  • Avoid input images with excessive noise, compression artifacts, or ambiguous foreground/background separation
  • The Preview version prioritizes speed over maximum quality; for final production, further refinement may be necessary
  • Prompt engineering can influence the style and mood of the generated video; descriptive prompts yield more controlled results
  • Iterative testing is recommended to fine-tune motion dynamics and visual effects
  • Be mindful of GPU memory requirements, especially when processing high-resolution images

Tips & Tricks

  • Start with clean, high-resolution images to maximize output quality
  • Use prompts that specify desired camera movement (e.g., “slow pan left,” “zoom in on subject”) for more predictable results
  • Experiment with atmospheric keywords such as “cinematic lighting,” “soft haze,” or “dramatic shadows” to enhance mood
  • Refine outputs by running multiple generations and selecting the best sequence; slight prompt adjustments can yield significantly different results
  • For advanced users, post-process generated videos with video editing tools to further enhance transitions or correct minor artifacts
  • If the output appears too static, try prompts that emphasize dynamic movement or environmental changes

Capabilities

  • Generates short, cinematic video sequences from a single input image
  • Preserves core details and composition of the original image while adding realistic motion
  • Supports a wide range of camera movements and atmospheric effects
  • Produces outputs suitable for concept visualization, storyboarding, and creative prototyping
  • Adapts well to various artistic styles and subject matter, from landscapes to portraits
  • Delivers fast generation times, enabling rapid iteration and experimentation

What Can I Use It For?

  • Professional concept visualization for film, animation, and advertising projects
  • Storyboarding and previsualization for creative teams
  • Rapid prototyping of motion ideas for photographers and digital artists
  • Enhancing social media content with cinematic video effects from still images
  • Educational demonstrations of image-to-video AI capabilities
  • Personal creative projects, such as turning travel photos into dynamic video memories
  • Industry-specific applications, including architecture walkthroughs and product showcases

Things to Be Aware Of

  • Some users report occasional artifacts or unnatural motion in highly detailed or complex scenes
  • The Preview version may not fully capture subtle lighting nuances compared to production-grade models
  • Generation speed is optimized, but output quality may require post-processing for professional use
  • GPU acceleration is recommended for best performance; CPU-only processing may be significantly slower
  • Consistency between frames is generally strong, but edge cases with ambiguous input images can result in flickering or jitter
  • Positive feedback highlights the model’s ease of use and impressive cinematic effects from simple inputs
  • Negative feedback centers on limitations in video length and occasional loss of fine image details

Limitations

  • Limited to short video sequences (typically 2-5 seconds); not suitable for long-form video generation
  • May struggle with highly complex scenes or images with ambiguous subject/background separation
  • Output quality, while strong for prototyping, may require additional refinement for final production use

Pricing

Pricing Type: Dynamic

720p resolution: duration * $0.10 per second from output video