Kling v2.6 Motion Control
Kling v2.6 Motion Control transfers full-body motion from a 3-30 second reference clip to a generated scene, capturing gestures, facial expressions, lip-sync, and camera movement with frame-accurate fidelity.
import { experimental_generateVideo as generateVideo } from 'ai';
const result = await generateVideo({ model: 'klingai/kling-v2.6-motion-control', prompt: 'A serene mountain lake at sunrise.'});Frequently Asked Questions
What format and duration should the reference video be?
The reference video should be 3-30 seconds long. Clearer subject visibility and stable framing produce more accurate motion transfer in the output.
Does Kling v2.6 Motion Control also transfer camera movement from the reference video?
Yes. Camera behavior (pan, push, pull, and rotation) in the reference clip replicates in the generated video, not just subject body motion.
How does it handle fast or complex movements like martial arts or dance?
The model reduces artifacts on fast, intricate motions. Hand articulation and high-speed body movements render with improved fidelity compared to earlier motion transfer approaches.
What is the maximum output duration?
Outputs can reach up to 30 seconds. This eliminates the need to stitch multiple short clips together for longer sequences.
Can the model transfer facial expressions and lip-sync from the reference video?
Yes. Facial expression tracking and lip-sync alignment transfer from the reference and apply to the generated subject.
Is a text prompt required alongside the reference video?
A text prompt is optional. Use it to describe the desired scene, subject, and styling for the output. Motion derives from the reference clip while the prompt defines what appears in the new video.