each::sense is in private beta.
Eachlabs | AI Workflows for app builders

KLING-V1

Kling v1 Standard Image to Video converts images into smooth, high-quality videos.

Avg Run Time: 270.000s

Model Slug: kling-v1-standard-image-to-video

Playground

Input

Enter a URL or choose a file from your computer.

Advanced Controls

Output

Example Result

Preview and download your result.

Unsupported conditions - pricing not available for this input format

API & SDK

Create a Prediction

Send a POST request to create a new prediction. This will return a prediction ID that you'll use to check the result. The request should include your model inputs and API key.

Get Prediction Result

Poll the prediction endpoint with the prediction ID until the result is ready. The API uses long-polling, so you'll need to repeatedly check until you receive a success status.

Readme

Table of Contents
Overview
Technical Specifications
Key Considerations
Tips & Tricks
Capabilities
What Can I Use It For?
Things to Be Aware Of
Limitations

Overview

Kling v1 Standard Image to Video generates short video sequences by transforming a static input image into dynamic motion guided by a descriptive text prompt. Kling v1 Standard Image to Video allows for customizable generation through various visual parameters including duration, aspect ratio, and auxiliary image inputs. Kling v1 Standard Image to Video is designed to create natural motion continuity between frames while preserving the original content structure of the image.

Technical Specifications

Kling v1 Standard Image to Video leverages advanced diffusion-based temporal modeling to generate consistent frame-to-frame motion.

Motion vectors are inferred from both prompt semantics and source image layout.

Designed to minimize flicker and artifacts by balancing global scene context with local pixel stability.

Supports frame-level interpolation and motion estimation between image pairs when tail_image_url is used.

Dynamic masking is internally applied to stabilize high-frequency regions unless overridden via static_mask_url.

Ensure the input image has a clear subject with minimal noise to maintain focus in motion rendering.

When using tail_image_url, select images with similar lighting and subject perspective to the main image_url for smoother transitions.

Keep prompts simple and descriptive; overly complex prompts can result in disjointed visuals.

Using a static mask (static_mask_url) can help maintain background or subject stability, depending on the use case.

Videos are currently limited to 5 or 10 seconds; longer durations are not supported.

Aspect ratio should match the subject orientation to avoid distortion.

Key Considerations

Input image quality directly affects the output. Low-resolution or overly compressed images may produce blurry or jittery results.

Prompts should be focused on motion, mood, or transformation. Avoid cluttering the prompt with scene descriptions already present in the image.

If both tail_image_url and static_mask_url are provided, the model prioritizes motion blending and overrides internal motion smoothing logic.

Videos are not audio-synced and contain no sound.


Legal Information for Kling v1 Standard Image to Video

By using this Kling v1 Standard Image to Video, you agree to:

Tips & Tricks

prompt
Use clear, action-oriented phrases (e.g., “a woman turning around slowly”, “clouds drifting across the sky”). Avoid abstract or poetic language.

cfg_scale
Controls adherence to the prompt.

  • Recommended value: 0.7
  • Lower values (0.3–0.5): more freedom, creative outputs.
  • Higher values (0.8–1): stricter adherence to prompt, but risk of less natural motion.

duration

  • Options: 5 or 10 seconds
  • Shorter durations result in more focused, stable animations.
  • For complex prompts, use 10 seconds to allow the model more frames to interpret motion.

aspect_ratio

  • Options: 16:9, 9:16, 1:1
  • Match subject framing:
    • 16:9 for landscape
    • 9:16 for portraits
    • 1:1 for centered subjects

image_url
Use high-quality images with the subject in the center. Plain or soft backgrounds produce cleaner animation.

tail_image_url
Adds dynamic ending. Use when transitioning between scenes or actions. Should visually align with the main image.

static_mask_url
Use this if part of the image should remain static. Ideal for keeping the background unchanged while animating the foreground.

negative_prompt
Use to exclude unwanted elements (e.g., “blurry, distorted, extra limbs”).

Capabilities

Transforms static images into short animated sequences.

Allows dynamic motion customization via textual descriptions.

Supports motion continuity between two input images.

Enables foreground/background isolation through masking.

Generates content with consistent subject focus and lighting retention.

What Can I Use It For?

Creating animated portraits or character loops.

Enhancing still images with realistic motion for digital content.

Producing short, looping visual stories for creative visuals.

Visualizing mood or atmosphere changes (e.g., lighting shifts, subtle motion).

Crafting seamless visual transitions between two related scenes.

Things to Be Aware Of

Animate a photograph of a person with a prompt like:
"a person smiling and tilting their head"

Combine two images (main and tail) with:

  • image_url: A person standing still
  • tail_image_url: Same person starting to walk
  • Prompt: "the person begins to walk forward"

Use static_mask_url to keep a building steady while animating the sky:

  • Prompt: "clouds slowly moving"
  • static_mask_url: mask over the building

Limitations

Limited to 5 or 10 seconds of output.

Model may struggle with complex or overlapping motion instructions.

Background artifacts may appear when subject edges are unclear.

Does not support facial lip-sync or precise expression control.

No support for audio integration.

Output Format: MP4

Pricing

Pricing Type: Dynamic

What this rule does

Pricing Rules

DurationPrice
5$0.14
10$0.28