POST
javascript
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 const axios = require('axios'); const api_key = "YOUR API-KEY"; const url = "https://api.segmind.com/v1/sync.so-lipsync-2-pro"; const data = { "video_url": "https://segmind-resources.s3.amazonaws.com/output/a741b039-226c-43c2-9bd0-c301f058d314-UntitledVideo-ezgif.com-crop-video.mp4", "audio_url": "https://segmind-resources.s3.amazonaws.com/output/80e96316-7e75-4733-b80c-049a0a6787cb-c9f17960-96b5-4119-8b7e-4ae0c9f21e2f-audio-AudioTrimmer.com-AudioTrimmer.com.mp3", "sync_mode": "loop", "temperature": 0.5, "auto_active_speaker_detection": true, "occlusion_detection_enabled": false }; (async function() { try { const response = await axios.post(url, data, { headers: { 'x-api-key': api_key } }); console.log(response.data); } catch (error) { console.error('Error:', error.response.data); } })();
RESPONSE
image/jpeg
HTTP Response Codes
200 - OKImage Generated
401 - UnauthorizedUser authentication failed
404 - Not FoundThe requested URL does not exist
405 - Method Not AllowedThe requested HTTP method is not allowed
406 - Not AcceptableNot enough credits
500 - Server ErrorServer had some issue with processing

Attributes


video_urlstr *

Provides the video URL for synchronization. Use high-quality links for best results.


audio_urlstr *

Provides the audio URL for synchronization. Use clear audio files for precision.


sync_modeenum:str ( default: loop )

Manages video-audio mismatch. Use 'loop' for repetitive audio, 'cut_off' for trimming.

Allowed values:


temperaturefloat ( default: 0.5 )

Controls expression in lip sync. Use 0.3 for calm, 0.8 for dynamic expressions.

min : 0,

max : 1


auto_active_speaker_detectionboolean ( default: true )

Detects and syncs active speaker automatically. Enable for multi-speaker scenarios.


occlusion_detection_enabledboolean ( default: 1 )

Detects occlusion, slowing generation. Disable for faster processing.

To keep track of your credit usage, you can inspect the response headers of each API call. The x-remaining-credits property will indicate the number of remaining credits in your account. Ensure you monitor this value to avoid any disruptions in your API usage.

Lipsync-2-Pro: AI Video Lip Synchronization Model

What is Lipsync-2-Pro?

Lipsync-2-Pro is an advanced AI model developed by Sync Labs that creates hyper-realistic lip synchronization videos. The AI model can effectively work across different video formats to edit dialogues within a video, while preserving facial expressions and even minute details with painstaking accuracy. Its 4K resolution, powered by diffusion-based super-resolution, gives speed in rendering video outputs with naturalistic results that don't need additional work in terms of training speakers in different languages or needing further refinement. Lipsync-2-Pro is a boon for film studios, content creators, and digital artists for its ability to create professional-level, perfectly synced videos.

Key Features Lipsync-2-Pro

  • It creates instant, high-quality lip synchronization videos for 4K videos
  • It effectively preserves facial details, even the minute details like teeth, freckles, and facial hair
  • It offers multi-language support with natural mouth movements
  • It seamlessly works with live-action, 3D animation, and AI-generated content
  • It can automatically detect the active speakers for multi-person scenes
  • It supports multiple sync modes to handle multiple audio-video scenarios
  • It includes adjustable expression control through temperature settings

Best Use Cases

  • It is used for film and TV post-production dubbing
  • It is ideal for podcast video content localization
  • It supports gaming cutscene dialogue modifications
  • It is excellent for educational content translations
  • It can be utilized for corporate training video updates
  • It will work flawlessly for live streaming content creation
  • It can work with virtual character animation
  • It is a high-end tool for multi-language marketing content

Prompt Tips and Output Quality

For optimal results -

  • Provide high-quality video and audio source files
  • Choose appropriate sync modes based on content -
    • Use 'loop' for repetitive audio sections
    • Select 'cut_off' for precise timing requirements
    • Enable 'bounce' for seamless continuous dialogue
  • Adjust temperature settings strategically:
    • Lower (0.3) for subtle, professional presentations
    • Higher (0.8) for dynamic, expressive content
  • Enable auto-active speaker detection for group scenes
  • Consider disabling occlusion detection for faster processing when precision isn't critical

FAQs

Q: How does Lipsync-2-Pro handle different languages? A: The AI model automatically adapts to any language to create the speaker's natural mouth movements without needing language-specific training.

Q: What video formats are supported? A: Lipsync-2-Pro works with multiple formats, which include live-action footage, 3D animations, and AI-generated videos up to 4K resolution.

Q: Do I need to train the model for different speakers? A: The huge advantage of the model is that it works instantly without speaker-specific training or fine-tuning.

Q: How can I optimize processing speed? A: You can optimize the processing speed by disabling occlusion detection for faster processing and ensuring clean audio input for best results.

Q: What's the recommended temperature setting? A: It is recommended to start with the default 0.5 setting; adjust lower (0.3) for subtle movements or higher (0.8) for more expressive results based on your content needs.

Q: Can it handle multiple speakers in one scene? A: Yes, you can enable the auto-active speaker detection for flawless synchronization in multi-speaker videos.