Unified character animation and replacement. Provide one character image and a reference video. Wan Animate reproduces expressions, body motion, and scene lighting to create coherent character videos or to replace a person in the scene.
A single design supports two tasks: animate a character from a performance, or replace the person in a video with your character while keeping motion and lighting consistent with the scene.
Give a character image and a reference video. The system reproduces facial expressions and full-body motion with clear timing and structure.
Place the animated character into the original video. Lighting and color tone are matched so the character fits the scene.
The character’s features remain stable across frames while expressions and motion vary with the performance.
Spatially aligned skeleton signals guide body motion, keeping timing stable from frame to frame.
Implicit features extracted from the source image help reproduce expressions with fine detail.
An auxiliary relighting adapter applies scene lighting and color tone without drifting the character’s identity.
Unified input design separates reference conditions from rendered regions.
Interact with the hosted demo to see character animation and replacement in action.
Follow a step-by-step guide to set up the model, prepare inputs, and run animation or replacement jobs on your machine.
The guide covers prerequisites, environment setup, downloading weights, and running the first example.
Produce character clips for short videos or prototypes. Keep expression, motion, and lighting consistent across shots.
Test staging and timing by animating stand-in characters from performance recordings.
Demonstrate motion mapping, face reenactment, and relighting concepts in a single project.
Study controllability and identity preservation with a unified interface for two related tasks.
Create character-driven variations while matching scene tone. Keep brand characters consistent across edits.
Build simple pipelines around a single image and a reference video to generate drafts quickly.
Inputs separate reference conditions from target regions to render. This single design covers both animation and replacement without extra tooling.
Body motion is guided by skeleton signals aligned in space. Facial reenactment uses features taken from the source image to keep identity stable while changing expression.
For replacement, an auxiliary adapter applies the scene’s lighting and color tone. Appearance stays consistent while the character fits the scene.
Outputs are controllable through skeleton strength, crop choices, and source image selection. Small changes in inputs help refine timing and expression.
At minimum, one high-quality character image and one reference video. A clear face and a stable pose in the image help the system reproduce expressions.
Yes. Choose the replacement task and provide the same inputs. The relighting adapter matches scene lighting and tone while keeping the character’s look.
Use a steady reference with clear body movement. Skeleton guidance maintains timing, and trimming shaky segments can improve results.
Yes. The model reads identity cues from the source image and maintains them across frames while expressions and motion follow the reference.