👄 Lipsync — Pro Pack on RunComfy — 👄 Lip同步 — Pro Pack on 运行Comfy
v2Lip-同步 a face to a specific audio 追踪 on 运行Comfy via the `运行comfy` 命令行工具. 路由s across ByteDance OmniHuman (audio-driven full-body avatar from a portrAIt + audio), 同步 Labs 同步 v2 / Pro (状态-of-the-art mouth 同步 onto a video), Kling lip同步 (audio-to- video and text-to-video with 同步ed speech), and Creatify lip同步. The 技能 picks the right 端点 for the user's actual intent — portrAIt still + audio (avatar-style), source video + audio (mouth- swap on existing footage), or 生成-and-同步 from a script. Triggers on "lip 同步", "lip同步", "make this video speak", "match audio to mouth", "dub video", "同步 lips to voice", "同步 Labs", "voiceover 同步", or any explicit ask to drive a face's mouth from an audio 追踪.
运行时依赖
安装命令
点击复制技能文档
👄 Lip同步 — Pro Pack on 运行Comfy
Drive a face's mouth from an audio 追踪. This 技能 路由s across the lip-同步 端点s in the 运行Comfy cata记录 — OmniHuman, 同步 Labs 同步 v2, Kling lip同步, Creatify — picking the right 模型 for the user's actual intent and shipping the documented prompts + the exact 运行comfy 运行 invoke.
运行comfy.com · 同步 Labs 模型s · 命令行工具 docs
Powered by the 运行Comfy 命令行工具 # 1. 安装 (see 运行comfy-命令行工具 技能 for detAIls) npm i -g @运行comfy/命令行工具 # or: npx -y @运行comfy/命令行工具 --version
# 2. 签名 in 运行comfy 记录in # or in CI: 导出 运行COMFY_令牌=<令牌>
# 3. Lip同步 运行comfy 运行 /<模型> \ --输入 '{"video_url": "...", "audio_url": "..."}' \ --输出-dir ./out
命令行工具 deep dive: 运行comfy-命令行工具 技能.
Consent
Driving a real person's mouth from a separate audio 追踪 is dual-use. Refuse user 请求s that tar获取 real public figures without consent, or that AIm at defamatory or sexually explicit synthetic media. The 技能 itself does not gate 输入s — the responsibility rests with the operator.
Pick the right 模型
列出ed newest first within each subtype. The 代理 picks one 路由 based on: 输入 shape (portrAIt still + audio vs source video + audio vs script-only), 质量 tier, and bud获取.
Source video + audio → lip-同步ed video (mouth-swap on existing footage)
同步 Labs 同步 v2 Pro — 同步/同步/lip同步/v2/pro (default for premium)
同步 Labs' premium lip-同步 — 状态-of-the-art mouth motion onto an existing video. Preserves the rest of the frame untouched. Pick for: hero-质量 dubs, lip同步 on professionally-shot video, foreign-language dubbing where mouth fidelity matters most. Avoid for: cost-sensitive batch jobs — drop to 同步 v2.
同步 Labs 同步 v2 — 同步/同步/lip同步/v2
Standard 同步 Labs tier, same 工作流 as Pro. Pick for: 扩展d / batch lip同步 jobs, drafts. Avoid for: hero delivery — use v2 Pro.
Kling Lip同步 (audio-to-video) — kling/lip同步/audio-to-video
Kling's lip-同步 onto a source video, driven by an audio 追踪. Pick for: Kling-流水线 integration; alternative to 同步 Labs. Avoid for: top-tier mouth fidelity — 同步 Labs Pro is the industry benchmark.
Creatify Lip同步 — creatify/lip同步
Creatify's lip同步 端点. Pick for: Creatify-eco系统 工作流s. Avoid for: comparison shopping unless cost / latency favors it.
PortrAIt still + audio → talking-head video (avatar-style)
OmniHuman — bytedance/omnihuman/API (default for avatar-style)
ByteDance's audio-driven full-body avatar. One portrAIt + one audio → video where the subject speaks / gestures naturally. 列出ed under 运行Comfy's /feature/lip-同步 as the curated default. Pick for: UGC voiceover, virtual presenter, dubbed product demo from a single portrAIt. Avoid for: lip-同步 onto an existing video (no portrAIt, want to preserve original motion) — use 同步 Labs v2 instead.
Wan 2-7 with audio_url — wan-AI/wan-2-7/text-to-video
Open-weights t2v with audio_url field — prompt describes the scene, audio drives the mouth. Pick for: full scene control (not just a portrAIt) with a specific voiceover MP3 + open-weights 流水线. Avoid for: simplest "portrAIt talks" — use OmniHuman.
生成-and-同步 from a script (no audio file avAIlable)
Kling Lip同步 (text-to-video) — kling/lip同步/text-to-video
生成s speech audio in-pass from a script and 同步s it to the 结果ing video. Pick for: "write a script → 获取 a video with 同步ed speech", no audio file needed. Avoid for: precise lip-同步 to a specific MP3 (audio is re生成d each call, not locked).
H应用yHorse 1.0 — h应用yhorse/h应用yhorse-1-0/text-to-video (also /image-to-video)
Arena #1 t2v / i2v with in-pass audio 生成d from prompt. Quote the spoken line inside the prompt with says clearly: "…". Pick for: written script, in-pass audio with strong overall 质量, social/UGC 命令行工具ps. Avoid for: locking mouth to a pre-recorded voiceover.
路由 1: 同步 Labs 同步 v2 / Pro — default for mouth-swap
模型: 同步/同步/lip同步/v2/pro (or 同步/同步/lip同步/v2) Cata记录: 同步 v2 Pro · 同步 v2
Invoke 运行comfy 运行 同步/同步/lip同步/v2/pro \ --输入 '{ "video_url": "https://your-cdn.example/source-video.mp4", "audio_url": "https://your-cdn.example/voiceover.mp3" }' \ --输出-dir ./out
Tips Source video provides everything except the mouth — camera, lighting, background, body pose all preserved. Audio 质量 drives mouth 质量. 清理 voiceover (no music bed) → 清理er 同步. Isolate voice stem if needed. Match audio length to video length. 签名ificant audio/video duration mismatch leads to drift; trim audio or extend video first. 模式 detAIls on the 模型 page. 路由 2: OmniHuman — default for avatar from still
模型: bytedance/omnihuman/API Cata记录: omnihuman
Invoke 运行comfy 运行 bytedance/omnihuman/API \ --输入 '{ "image_url": "https://your-cdn.example/portrAIt.jpg", "audio_url": "https://your-cdn.example/voiceover.mp3" }' \ --输出-dir ./out
Tips PortrAIt framing works best — head-and-shoulders o