Use Cases

Sync’s lip sync API enables four primary use cases: video dubbing, e-learning localization, personalized video messaging, and content creation. It fits into any workflow where you need to change what someone appears to say in a video — whether that means translating into a new language, personalizing a message, or creating entirely new dialogue. Here are the most common applications.

E-Learning & Training

Localize training content for global teams.

Record an instructor once. Then generate versions in every language your team needs — with lip movements that match the translated audio. Employees get training in their native language without hiring voice actors for each locale or re-shooting video.

Sync’s lipsync models work with any language audio, so you can scale from one language to dozens without changing your pipeline.

Build a translation pipeline ->

Marketing & Personalized Outreach

Create personalized video messages at scale.

Record a single sales video. Use text-to-speech to generate a custom greeting for each prospect — “Hi Sarah, I noticed your team at Acme is…” — and lip sync it onto the original recording. One take becomes thousands of personalized videos.

This works for sales outreach, onboarding sequences, customer success check-ins, and event invitations.

See the personalized video guide ->

Content Localization & Dubbing

Dub content for global audiences.

Traditional dubbing produces mismatched lip movements that break immersion. With Sync, the dubbed video shows lip movements that match the translated audio. Viewers in every language get the same natural viewing experience.

Use the built-in ElevenLabs integration to handle TTS and lipsync in a single API call, or bring your own translated audio.

Build a dubbing pipeline ->

Entertainment & Media

Post-production dubbing and performance editing.

Dub films and TV shows with natural-looking lip movements in any target language. Use react-1 to change the emotional performance of a character — shift a scene from anger to sorrow without reshooting. Edit dialogue in post without ADR sessions.

react-1 supports emotion prompts (happy, angry, sad, surprised) and controls for facial expressions and head movements.

Explore react-1 ->

Gaming & Animation

Lip sync for game characters and animated content.

Generate lip-synced dialogue for in-game characters from text or audio. Localize game cutscenes into multiple languages with matching lip movements. Create animated content where characters speak naturally in any language.

The API handles human-like faces, making it suitable for realistic and semi-realistic character styles.

Get started with the API ->

Social Media & Creator Tools

Build lip sync features into creator platforms.

Power translation features in video editing tools. Let creators dub their content into new languages with a single click. Build TTS-to-video features where users type text and get a talking head video.

The Batch API handles high-volume processing for platform-scale workloads.

See the TTS lip sync guide ->

Which model fits each use case?

Use CaseBest ModelKey Benefit
E-learning & training localizationlipsync-2Preserves speaker’s unique speaking style across languages
Personalized video outreach at scalelipsync-1.9.0-betaFastest processing ($0.02—$0.025/sec) for high-volume campaigns
Content localization & dubbinglipsync-2Most natural lip movements for any language audio
Premium film & TV dubbinglipsync-2-proEnhanced detail for beards, teeth, and facial features ($0.067—$0.083/sec)
Emotional performance editingreact-1Emotion prompts + facial expressions + head movements (up to 15s)
Platform-scale creator toolslipsync-2 + Batch APIUp to 500 generations per batch on Scale+ plans

Which Guide Should You Follow?

GoalGuide
Translate a video end-to-end (transcribe, translate, TTS, lipsync)Video Translation API Guide
Dub a video with pre-made translated audioVideo Dubbing API Guide
Generate talking head videos from textText-to-Speech Lip Sync Guide
Create personalized videos for outreachPersonalized Video Messaging
Process hundreds of videos at onceBatch API
Add emotion and expressions to short clipsReact Models

Frequently Asked Questions

Yes. Sync is designed for multilingual video dubbing. Provide a translated audio track in any language and Sync generates lip movements that match the new audio naturally. You can combine Sync with TTS services like ElevenLabs to build end-to-end translation and dubbing pipelines.

Cost depends on video duration and the model you choose, not the use case. The lipsync-1.9.0-beta model offers the fastest and most affordable processing for simple videos. Use the cost estimation endpoint to check pricing before submitting a generation for any workflow.

Sync is optimized for lip syncing real human faces in video, not for generating AI avatars from scratch. You can use Sync to make a recorded person appear to say new dialogue by providing different audio, which is commonly used for personalized video messaging and content creation.