LipsyncX
Video-to-Video model

Sync Lipsync 2.0

Balanced quality and speed for general lip‑sync dubbing.

Zero‑shot video‑to‑video lip sync that preserves a speaker’s style while matching new audio. Built for editing dialogue or dubbing across live‑action, animation, and AI‑generated humans without retraining.

Best for: Creator videos
Inputs: Video + Audio
Outputs: Video

What this model is best at

Short answer: Zero‑shot video‑to‑video lip sync that preserves a speaker’s style while matching new audio. Built for editing dialogue or dubbing across live‑action, animation, and AI‑generated humans without retraining.

Use this workspace to preview the model, compare example output, and start creating with the recommended workflow for this model.

Highlight 1

Zero‑shot editing with no actor training required.

Highlight 2

Preserves unique speaking style and cadence.

Highlight 3

Works with live‑action, animation, and AI‑generated characters.

Video-to-Video

Sync Lipsync 2.0 workspace

Start from the built-in workflow below, then tune the model inside the standard LipsyncX creation surface.

Popular model
Creator videos · Marketing clips · Standard dubbing

1. Upload photo

1. Choose a face

Step 1/4

Choose a face

Follow the next step to keep building your video.

UGC ad re‑dub

Swap a new hook while preserving the original footage.

Original
UGC ad re‑dub original
Synced
UGC ad re‑dub generated

Popular use cases

Use case 1

UGC variations

Rotate new scripts without reshoots.

Use case 2

Explainers

Keep visuals, change narration fast.

Use case 3

Creator content

Ship updates with the same host.

Quick specs

Primary use
Video‑to‑video lip sync
Inputs
Source video + target audio
Output
Synced video
Best strength
Balanced quality and speed

Best practices

Use clear, studio‑quality audio for the target voice.
Keep the face large and well‑lit for best mouth detail.
Match the emotional tone of the original performance.

FAQ

Do I need to train on the speaker first?

No. Lipsync‑2 is zero‑shot, so it can edit any speaker without training.

What kinds of footage does it support?

It works on live‑action video, animation, and AI‑generated humans.

What inputs are required?

Provide a source video plus target audio (or a script + voice) via the API.

Ready to try Sync Lipsync 2.0?

Use the built-in workspace to test prompts, compare outputs, and see how this model fits your content workflow.