sync-3

The most powerful lip sync model ever built.

Quick reference

Model namesync-3
StatusDefault model for all users
API endpointPOST /v2/generate
Languages95+
Face resolution4K native output
Pricing (at 25 fps)$0.107 – $0.133/sec
Available inAPI, Studio, Adobe Premiere Plugin

What’s new

Previous models processed video in small, independent snippets. sync-3 takes a fundamentally different approach — it builds a global understanding of a person across an entire shot, generating all frames at once rather than stitching together isolated segments.

The result is a generational shift in consistency and realism.

CapabilityWhat changed
Close-ups & partial facesTight close-ups, cropped frames, and partially obscured faces are handled natively — cinematic and editorial content that was previously off-limits.
Extreme anglesProfile shots, over-the-shoulder angles, and non-frontal lip positions that broke earlier models are now handled with confidence.
Obstruction detectionHands, microphones, scarves — sync-3 detects obstructions automatically and generates around them without manual intervention.
Style & emotion preservationPreserves the original speaker’s cadence and emotional expression. Silent lips can be opened naturally. The output isn’t just accurate — it feels like the person actually said it.

How it works

sync-3 generates from a larger spatial window than any previous model, giving it a much wider field of view around the face. The model can reason about what to generate and what to preserve because it has enough context to understand the full scene.

This means fewer retakes, fewer manual fixes, and the best generation possible in any scenario.

Options not available

The following generation options from previous models are not applicable to sync-3. Their capabilities are either built into the model or handled differently by the new architecture.

OptionWhy it’s not needed
temperaturesync-3 manages expressiveness natively — no manual tuning required.
reasoning_enabledFrame analysis and correction for artifacts, occlusions, and extreme poses is built into the model.
occlusion_detection_enabledObstruction detection is automatic — sync-3 detects and generates around obstructions without manual intervention.

Integration

sync-3 is available through the standard Sync API. Pass sync-3 as the model parameter in your generation request.

1from sync import Sync
2from sync.common import Audio, Video
3
4sync = Sync()
5
6response = sync.generations.create(
7 input=[
8 Video(url="https://assets.sync.so/docs/example-video.mp4"),
9 Audio(url="https://assets.sync.so/docs/example-audio.wav")
10 ],
11 model="sync-3"
12)

Need to get set up first? See the Quickstart guide for API key creation and SDK installation.

Works with

Also available via ComfyUI and MCP Server for AI-assisted workflows.

How sync-3 compares

sync-3lipsync-2-prolipsync-2
Face resolution4K native512×512 (enhanced detail)512×512
Close-ups & partial facesNativeLimitedLimited
Extreme anglesNativeLimitedLimited
Obstruction detectionAutomaticManual opt-inManual opt-in
Emotion preservationFull cadence & expressionPreserves speaking stylePreserves speaking style
Silent lip openingYesNoNo
ProcessingFull-shot (all frames at once)2-second independent chunks2-second independent chunks
Pricing (at 25 fps)$0.107 – $0.133/sec$0.067 – $0.083/sec$0.04 – $0.05/sec

sync-3 is the best choice for production-grade video — especially content with close-ups, complex angles, or obstructions. For simpler videos where cost is a priority, lipsync-2 or lipsync-2-pro remain excellent options.

FAQs

Change the model parameter from your current model (e.g. lipsync-2) to sync-3. The rest of the request schema is unchanged — no other code changes required.

No. sync-3 detects obstructions automatically. You no longer need to set occlusion_detection_enabled in your request options.

sync-3 supports 95+ languages — the same broad language coverage as previous models. It’s designed for global dubbing at native quality.

Yes. sync-3 works with the Batch API the same way as other models — just set model to sync-3 in each job.