1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
const axios = require('axios');
const api_key = "YOUR API-KEY";
const url = "https://api.segmind.com/v1/kling-o1-image-to-video";
const data = {
"prompt": "A seamless transformation of a man driving a vehicle and crashing",
"start_image_url": "https://segmind-resources.s3.amazonaws.com/output/d472b3e1-2b68-406d-9d12-a690c3da4045-seedance_1.5_input.webp",
"end_image_url": "https://segmind-resources.s3.amazonaws.com/output/dbc484a8-bcda-4b9c-97f6-d89dca2f3815-output-1767723467082.png",
"duration": 5
};
(async function() {
try {
const response = await axios.post(url, data, { headers: { 'x-api-key': api_key } });
console.log(response.data);
} catch (error) {
console.error('Error:', error.response.data);
}
})();Describe the video transformation or animation between start and end images
URL of the starting frame image
URL of the ending frame image (optional)
Video duration in seconds
Allowed values:
To keep track of your credit usage, you can inspect the response headers of each API call. The x-remaining-credits property will indicate the number of remaining credits in your account. Ensure you monitor this value to avoid any disruptions in your API usage.
Edited by Segmind Team on January 26, 2026.
Kling Omni Video O1, by Kuaishou, is a cutting-edge image-to-video AI model that impressively brings still images to life by transforming them into cinematic animations while sticking to principles of physics for realistic output. It goes beyond conventional interpolation tools to effectively understand the nuances of scene dynamics to generate dynamic, realistic movements that appear authentic by 'respecting' the physics and spatial interactions. Kling Omni Video O1 is built with the Multi-modal Visual Language (MVL) technology, hence it perfectly upholds subject consistency, ensuring character identity, props, color tones, and lighting across the entire video sequence. All these aspects make this model an asset to developers who can harness its creative power through a REST API to produce professional-quality videos at scale and even within a limited timeline.
How does Kling Omni Video O1 differ from standard video interpolation?
Kling Omni Video O1 uses MVL technology to understand scene context and generate physics-based motion, thereby creating new frames with natural movement rather than morphing pixels; this aspect makes it significantly better than frame interpolation tools that simply blend between images.
Can I control camera movement separately from subject motion?
To achieve multi-layered control in camera movement, include camera directives in your prompt: e.g., "camera slowly zooms in," "tracking shot following subject", etc., alongside subject actions.
What image formats and resolutions work best?
Kling Omni Video O1 accepts standard web image formats via URL. To achieve the high-end results, use images with clear subjects, good contrast, and resolution of at least 1024px on the longest side.
Is an end image required for good results?
No, the model works well with only a start image and text prompt. You may add an end image for precise transformation control or specific final states.
How do I optimize for faster processing?
Select the 5-second duration option and ensure your start image URL is hosted on fast, reliable infrastructure to minimize upload time.
What happens to small details in the original image?
Kling Omni Video O1's subject consistency algorithms work efficiently to preserve fine details, textures, and color information throughout the animation; though extreme detail may soften slightly during rapid motion.