Home AIツール AIガイド AIモデル AIクリエイター 🛒 購入 始める
🎬 KLING AI ⏱ 4 min read 🎬 モーションコントロール

Motion Control — Technical Guide

写真とモーションビデオをアップロードします — あなたのキャラクターは新しいAIビデオで正確に同じ動き、ダンス、またはジェスチャーを行います

🎬

Motion Control

写真とモーションビデオをアップロードします — あなたのキャラクターは新しいAIビデオで正確に同じ動き、ダンス、またはジェスチャーを行います
ダンスビデオを見て、あなたのキャラクターが同じ動きをできたらいいなと思ったことはありませんか?それがまさにモーションコントロールの機能です。

あなたのキャラクターの写真と、誰かが動きを行っているビデオ — ダンス、歩行、手を振る、フィットネスルーチンを行う — をアップロードすると、AIはあなたのキャラクターがその正確な動きを行う新しいビデオを生成します。あなたのキャラクターは顔、衣装、アイデンティティを保持しますが、今では参照ビデオの人物と同じように動いています。

2つのモードで結果を制御できます:カメラフォローは参照ビデオと同じカメラアングルを保持し、イメージマッチはあなたのキャラクターの元のポーズの方向を保持し、プロンプトを通じてカメラの動きを追加できます。

トレンドの振り付けに合わせてAIキャラクターをダンスさせたり、シーンを歩かせたり、手話を行ったり、エクササイズを示したり、撮影またはオンラインで見つけた任意の身体的動きを演じたりするのに最適です。
✦ Best Results Tips
🧍 全身を見せる
あなたのキャラクターの写真とモーションビデオは、全身を明確に示す必要があります — 頭からつま先まで、切り取らないでください。AIは動きを正確に転送するために、すべての手足を見る必要があります。
🎬 1つの連続ショット
モーションビデオは、1つの途切れのないテイクでなければなりません。カットなし、シーン変更なし、カメラ切り替えなし。動きの開始から終了までの滑らかなクリップだけです。
🤚 動きを安定させる
遅くて穏やかな動きが最良の結果を生み出します。速いまたは不規則な動作は、転送中に失われたり歪んだりする傾向があります。
👤 1人のみ
はっきりと見える1人だけのモーションビデオを使用してください。複数の人がいる場合、AIは最も大きな人を選びます — それがあなたの意図した人物でない可能性があります。
📐 フレーミングを合わせる
あなたのキャラクターの写真が全身ショットの場合、全身のモーションビデオを使用してください。半身の写真の場合は、半身のモーションビデオを使用してください。フレーミングが一致しないと、不自然な結果が生じます。
🏃 リアルな人間の動きが最適
あなたが望むアクションを実行しているリアルな人を撮影してください。アニメーションやCGIの参照ビデオは、リアルに撮影された人間の動きほどうまく転送されません。

Motion Control — Available Models

Kling v2.6
Default
kling-v2-6
Stable motion transfer from reference video.
Kling v3.0
kling-v3
Higher quality motion control.
📥 You Give
🖼️Character Image 🎬Motion Reference Video 📝Text Prompt (optional)
AI Magic
klingai
🎬 You Get
🎬 Video
Quality modes
Standard
Professional
Orientation modes
Camera Follow (up to 30s)
Image Match (up to 10s)
🔇
Keep audio
Preserve original sound
⏱️
30s
Max duration
🎥
MP4/MOV, 3-30s, max 100MB
Video input
🖼️
JPG/PNG, max 10MB
Image input

💰 Motion Control — Pricing

Estimated cost
Failed jobs are automatically refunded

Kling VIDEO 3.0 Motion Control

The newly released Kling VIDEO 3.0 Motion Control builds upon the Motion Control introduced in VIDEO 2.6, delivering key capability upgrades. VIDEO 3.0 Motion Control enhances facial consistency across scenarios, ensuring stable facial features and smooth expressions even in complex, multi-angle, long-duration motions. This upgrade expands Motion Control into cinematic performance, high-precision motion capture, and diverse entertainment scenarios, delivering more powerful and reliable video generation.

Consistent Facial Identity from Any Angle

Reference Image
Element
Output with Element Binding
Output without Element Binding
Reference
Reference Image
Element
Output with Element Binding
Output without Element Binding
Reference

Complex Emotions, Faithfully Reproduced

Reference Image
Element
Output with Element Binding
Output without Element Binding
Reference
Element
Element
Element
Reference Image
Element
Output with Element Binding
Output without Element Binding
Reference
Element
Element
Element
Element

Face Occlusion, High-Fidelity Restoration

Reference Image
Element
Output with Element Binding
Output without Element Binding
Reference
Element
Element
Element

Consistent Facial Clarity Across Dynamic Framing

Reference Image
Element
Output with Element Binding
Output without Element Binding
Reference
Reference Image
Element
Output with Element Binding
Output without Element Binding
Reference

How to Achieve the Desired Outputs

1. The Motion Control Element Library only uses facial information for reference. It does not include clothing, hairstyle, makeup, or props. Therefore, we recommend uploading clear facial close-ups to ensure sufficient facial data.

2. Whether you upload images or videos, follow this core principle: Upload facial references that match the result you want to generate.

a) Head Turn Accuracy — To achieve more accurate head turns, upload: a front-facing view, side views (left and/or right).

b) Facial Expression Accuracy — To better match facial expressions (such as smiling), upload: a neutral front-facing image, a smiling front-facing image.

c) 360° Smiling Rotation — For a seamless 360° smiling rotation, upload: front-facing smile, left-profile smile, right-profile smile, upward-facing smile, downward-facing smile.

d) Complex Emotional Transitions with Head Movement — For complex emotional changes (e.g. happy to sad) combined with head turns, upload: a front-facing image, a smiling expression, a sad expression, side views (left or right).

e) If you need complex facial expressions while maintaining high identity accuracy, we strongly recommend uploading a video, which provides richer and more continuous facial information.

3. Edge Cases

— The first frame in Motion Control may contain multiple people, but only one element is supported; the system will select the person with the largest on-screen presence as the element. If the elements occupy similar portions of the frame, no element will be selected.

— If the element's face differs significantly from the face in the first frame, there is a small chance that facial quality may degrade — for example, when using a cat's face to reference a human.

Technical Requirements:

1. Ensure the character's entire body and head are clearly visible and not obstructed.

2. Upload a single character motion reference. For motion references with two or more characters, the motion of the character occupying the largest portion of the frame will be used for generation.

3. Real human actions are recommended, while certain stylized humanoid or humanoid body proportions can be recognized.

4. The action video must be a single continuous shot, with the character consistently visible in the frame. Please avoid cuts, shot changes, or camera movements; otherwise, the video may be truncated.

5. Avoid overly fast motions; steady, moderate movements yield the best results.

6. The short edge must be at least 340px, and the long edge must not exceed 3850px.

7. The supported duration for uploaded action videos is 3–30 seconds, and the generated video duration will match the length of the uploaded video. If the action is highly complex or performed at a very fast pace, there is a possibility that the generated result may be shorter than the original upload. This is because the model extracts only the valid and continuous action segments for generation. As long as a minimum of 3 seconds of usable continuous motion is extracted, the video can be generated. Please note that in such cases, the consumed Credits are non-refundable. We recommend adjusting the action difficulty and speed accordingly for optimal results.

Perfectly Synchronized Full-Body Motions

Image Reference
Motion Reference
Output
Reference

How to Achieve the Desired Outputs

1. Match the character's full-body/half-body in the image reference with the full-body/half-body in the motion reference.

2. Use a motion reference that features a wide range of motion, moderate speed, and minimal displacement.

3. For large motion references, ensure there is enough space in the image reference for the character to move freely.

Image Reference

Half Body
Half Body
Half Body
Half Body
Full body
Full body
Full body
Full body

Motion Reference

Half-Body
Full-Body
Full-Body

1. Ensure the character's entire body and head are clearly visible and not obstructed.

2. Upload a single character motion reference. For motion references with two or more characters, the motion of the character occupying the largest portion of the frame will be used for generation.

3. Real human actions are recommended, while certain stylized humanoid or humanoid body proportions can be recognized.

4. Avoid cuts and camera movements in the motion reference.

5. Avoid overly fast motions; steady, moderate movements yield the best results.

6. The short edge must be at least 340px, and the long edge must not exceed 3850px.

7. The duration range of the uploaded motion reference is from 3 to 30 seconds, in which the generated video length will align with the duration of the uploaded video. If motions are complex or fast-paced, there is a chance that the output may be shorter than the uploaded video duration, as the model can only extract the valid action duration for generation. The minimum extractable continuous action duration is 3 seconds. Please note that in such cases, the consumed credits cannot be refunded. It is recommended to adjust the complexity and speed of the actions accordingly.

Character Orientation

1. By default, the video will be generated through "Character Orientation Matches Video", and the character's movements, expressions, camera movements, and orientation will follow the motion reference. Other details can be controlled via prompts.

2. When you choose "Character Orientation Matches Image" to match the character orientation with the image reference, the character's movements and expressions will follow the motion reference, and the orientation will align with the character orientation in the reference image. Camera movements and other elements can be customised through prompts.

Character Orientation Matches Image

Camera Movement Showcase

Zoom In
Zoom Out
Camera Up
Camera Down
Fixed Position

🎬 モーションコントロール

モーションコントロールを試す