Use Cases
Sync’s lip sync API enables four primary use cases: video dubbing, e-learning localization, personalized video messaging, and content creation. It fits into any workflow where you need to change what someone appears to say in a video — whether that means translating into a new language, personalizing a message, or creating entirely new dialogue. Here are the most common applications.
Localize training content for global teams.
Record an instructor once. Then generate versions in every language your team needs — with lip movements that match the translated audio. Employees get training in their native language without hiring voice actors for each locale or re-shooting video.
Sync’s lipsync models work with any language audio, so you can scale from one language to dozens without changing your pipeline.
Create personalized video messages at scale.
Record a single sales video. Use text-to-speech to generate a custom greeting for each prospect — “Hi Sarah, I noticed your team at Acme is…” — and lip sync it onto the original recording. One take becomes thousands of personalized videos.
This works for sales outreach, onboarding sequences, customer success check-ins, and event invitations.
Dub content for global audiences.
Traditional dubbing produces mismatched lip movements that break immersion. With Sync, the dubbed video shows lip movements that match the translated audio. Viewers in every language get the same natural viewing experience.
Use the built-in ElevenLabs integration to handle TTS and lipsync in a single API call, or bring your own translated audio.
Post-production dubbing and performance editing.
Dub films and TV shows with natural-looking lip movements in any target language. Use react-1 to change the emotional performance of a character — shift a scene from anger to sorrow without reshooting. Edit dialogue in post without ADR sessions.
react-1 supports emotion prompts (happy, angry, sad, surprised) and controls for facial expressions and head movements.
Lip sync for game characters and animated content.
Generate lip-synced dialogue for in-game characters from text or audio. Localize game cutscenes into multiple languages with matching lip movements. Create animated content where characters speak naturally in any language.
The API handles human-like faces, making it suitable for realistic and semi-realistic character styles.
Build lip sync features into creator platforms.
Power translation features in video editing tools. Let creators dub their content into new languages with a single click. Build TTS-to-video features where users type text and get a talking head video.
The Batch API handles high-volume processing for platform-scale workloads.
Which model fits each use case?
Which Guide Should You Follow?
Frequently Asked Questions
Can Sync dub videos into other languages?
Yes. Sync is designed for multilingual video dubbing. Provide a translated audio track in any language and Sync generates lip movements that match the new audio naturally. You can combine Sync with TTS services like ElevenLabs to build end-to-end translation and dubbing pipelines.
Which use case is cheapest?
Cost depends on video duration and the model you choose, not the use case. The lipsync-1.9.0-beta model offers the fastest and most affordable processing for simple videos. Use the cost estimation endpoint to check pricing before submitting a generation for any workflow.
Can I use Sync for AI avatars?
Sync is optimized for lip syncing real human faces in video, not for generating AI avatars from scratch. You can use Sync to make a recorded person appear to say new dialogue by providing different audio, which is commonly used for personalized video messaging and content creation.

