DeskRex.ai

open α

テーマ

データベース

自動化

発見

サインイン

リサーチの結果の保存、レポートの作成、共有が行えます。

サインイン
データベースに戻る

📀 出典のデータベース: 無料のリップシンク対応生成AIアバター用ツール・OSSライブラリ

ツール・ライブラリ名

Rhubarb Lip Sync

🗓 更新日: 5/13/2025

公開中

項目

対応プラットフォームリップシンク機能の有無無料/OSSの区分備考・特徴関連する出典情報

対応プラットフォーム

Windows、Linux、Mac(コマンドラインツール), Windows, MacOS, Linux (コマンドラインツール), コマンドラインツール、Blender、Godot Engine対応, Windows, macOS, Linux, Blender, OpenToonz, Adobe After Effects, Spine, Vegas Pro, Windows、macOS、Linux, Windows, macOS, Linux
vozo.ai
note.com
github.com
blenderartists.org
godotengine.org
github.com
github.com
linkedin.com
github.com
github.com
github.com
x.com

リップシンク機能の有無

✅
vozo.ai
note.com
github.com
blenderartists.org
godotengine.org
github.com
github.com
linkedin.com
github.com
github.com
github.com
x.com

無料/OSSの区分

OSS(オープンソースソフトウェア), OSS, OSS(オープンソース), オープンソース(無料)
vozo.ai
note.com
github.com
blenderartists.org
godotengine.org
github.com
github.com
linkedin.com
github.com
github.com
github.com
x.com

備考・特徴

音声録音から2D口アニメーションを自動生成するコマンドラインツール。ゲームキャラクターのアニメーションに利用可能。OpenToonzやGodot Engineとの連携あり。高速リップシンク対応。, 音声録音から2D口アニメーションを自動生成。OpenToonzやGodot Engineとの連携あり。ゲームキャラクターのアニメーションに利用可能。, 音声録音から2D口アニメーションを自動生成。OpenToonzやGodot Engineとの連携が可能。高速リップシンク対応。コマンドラインツール。, 音声録音から2D口アニメーションを自動生成するコマンドラインツール。多様な出力形式に対応。BlenderやOpenToonz、Adobe After Effectsなど複数のツールと連携可能。高精度なリップシンクを実現。, 音声録音から2D口アニメーションを自動生成するコマンドラインツール。Adobe After Effects、Moho、OpenToonzなどのアニメーションソフトと連携可能。多彩な出力形式対応。英語音声に特化したPocketSphinxと音素認識方式を選択可能。, 音声録音から2D口アニメーション自動生成。OpenToonzやGodot Engineと連携可能。ゲームキャラクター向け。
vozo.ai
note.com
github.com
blenderartists.org
godotengine.org
github.com
github.com
linkedin.com
github.com
github.com
github.com
x.com
visionaire-studio.net

関連する出典情報

copy url
source logonote.com
AI Fusion
OVR Lipsync
NVIDIA Omniverse Audio2Face
Facial Action Coding System
VOCAL: Vowel and Consonant Layering for Expressive Animator-Centric Singing Animation
CEDEC 2022 機械学習によるリップシンクアニメーション自動生成技術とFINAL FANTASY Ⅶ REMAKE のアセットを訓練データとした実装事例
UXDevシリーズ
AI Fusionプレスリリース
uLipsync
hecomi
https://github.com/hecomi/uLipSync/releases/tag/v3.0.0
MetaCommunication
GREE Tech Conference2021
https://j.mp/VRSYT
VTechChallenge2023
VRSionUp!13
https://youtube.com/live/9achiaxHc-4
[インターン希望の方へ
copy url
source logogithub.com
HERE
HunyuanPortrait: Implicit Condition Control for Enhanced Portrait Animation
[Code
[Project
X-Portrait: Expressive Portrait Animation with Hierarchical Motion Attention
LivePortrait: Efficient Portrait Animation with Stitching and Retargeting Control
[Code
[Project
EMOPortraits: Emotion-enhanced Multimodal One-shot Head Avatars
[Code
[Project
Synergizing Motion and Appearance: Multi-Scale Compensatory Codebooks for Talking Head Video Generation
[Project
Audio-Visual Face Reenactment
[Code
[Project
Cross-identity Video Motion Retargeting with Joint Transformation and Synthesis
[Code
Implicit Identity Representation Conditioned Memory Compensation Network for Talking Head Video Generation
[Project
[Code
Depth-Aware Generative Adversarial Network for Talking Head Video Generation
[Code
[Project
Thin-Plate Spline Motion Model for Image Animation
[Code
StyleHEAT: One-Shot High-Resolution Editable Talking Face Generation via Pretrained StyleGAN
[Code
[Project
MegaPortraits: One-shot Megapixel Neural Head Avatars
[Project
Structure-Aware Motion Transfer with Deformable Anchor Model
[Code
StyleMask: Disentangling the Style Space of StyleGAN2 for Neural Face Reenactment
[Code
Controllable Radiance Fields for Dynamic Face Synthesis
Animatable 3D-Aware Face Image Generation for Video Avatars
[Project
Implicit Warping for Animation with Image Sets
[Project
HifiHead: One-Shot High Fidelity Neural Head Synthesis with 3D Control
Face Animation with Multiple Source Images
MetaPortrait: Identity-Preserving Talking Head Generation with Fast Personalized Adaptation
Compressing Video Calls using Synthetic Talking Heads
[Project
Finding Directions in GAN’s Latent Space for Neural Face Reenactment
[Project
[Code
Latent Image Animator: Learning to Animate Images via Latent Space Navigation
[Project
[Code
One-Shot Free-View Neural Talking-Head Synthesis for Video Conferencing
[Project
Sparse to Dense Motion Transfer for Face Image Animation
SAFA: Structure Aware Face Animation
[Code
Self-appearance-aided Differential Evolution for Motion Transfer
PIRenderer: Controllable Portrait Image Generation via Semantic Neural Rendering
[Code
FACEGAN: Facial Attribute Controllable rEenactment GAN
F3A-GAN: Facial Flow for Face Animation With Generative Adversarial Networks
FACIAL: Synthesizing Dynamic Talking Face with Implicit Attribute Learning
Motion Representations for Articulated Animation
[Code
HeadGAN: One-shot Neural Head Synthesis and Editing
[Project
Mesh Guided One-shot Face Reenactment Using Graph Convolutional Networks
[Code
MarioNETte: Few-shot Face Reenactment Preserving Identity of Unseen Targets
[Project
Learning Identity-Invariant Motion Representations for Cross-ID Face Reenactment
First order motion model for image animation
[Code
Few-Shot Adversarial Learning of Realistic Neural Talking Head models
[Code
Animating Arbitrary Objects via Deep Motion Transfer
[Code
[Project
Few-shot Video-to-Video Synthesis
[Code
[Project
ReenactGAN: Learning to Reenact Faces via Boundary Transfer
[Code
X2Face: A network for controlling face generation by using images, audio, and pose codes
[Code
[Project
Face2Face: Real-time face capture and reenactment of RGB videos
OmniHuman-1: Rethinking the Scaling-Up of One-Stage Conditioned Human Animation Models
[Project
Audio-visual Controlled Video Diffusion with Masked Selective State Spaces Modelling for Natural Talking Head Generation
[Project
Real3D-Portrait: One-shot Realistic 3D Talking Portrait Synthesis
[Project
[Code
Emote Portrait Alive - Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions
[Project
[Code
Style2Talker: High-Resolution Talking Head Generation with Emotion Style and Art Style
Say Anything with Any Style
Code
[Project
THQA: A Perceptual Quality Assessment Database for Talking Heads
[Code
Talk3D: High-Fidelity Talking Portrait Synthesis via Personalized 3D Generative Prior
[Code
[Project
EDTalk: Efficient Disentanglement for Emotional Talking Head Synthesis
[Code
[Project
AniPortrait: Audio-Driven Synthesis of Photorealistic Portrait Animations
[Code
FlowVQTalker: High-Quality Emotional Talking Face Generation through Normalizing Flow and Quantization
FaceChain-ImagineID: Freely Crafting High-Fidelity Diverse Talking Faces from Disentangled Audio
[Code
Hallo: Hierarchical Audio-Driven Visual Synthesis for Portrait Image Animation
[Code
EchoMimic: Lifelike Audio-Driven Portrait Animations through Editable Landmark Conditions
[Code
[Project
RealTalk: Real-time and Realistic Audio-driven Face Generation with 3D Facial Prior-guided Identity Alignment Network
Emotional Conversation: Empowering Talking Faces with Cohesive Expression, Gaze and Pose Generation
Make Your Actor Talk: Generalizable and High-Fidelity Lip Sync with Motion and Appearance Disentanglement
FD2Talk: Towards Generalized Talking Head Generation with Facial Decoupled Diffusion Model
ReSyncer: Rewiring Style-based Generator for Unified Audio-Visually Synced Facial Performer
Style-Preserving Lip Sync via Audio-Aware Style Reference
Loopy: Taming Audio-Driven Portrait Avatar with Long-Term Motion Dependency
[Project
DAWN: Dynamic Frame Avatar with Non-autoregressive Diffusion Framework for Talking Head Video Generation
[Project
[Code
EchoMimicV2: Towards Striking, Simplified, and Semi-Body Human Animation
[Code
[Project
Latent Diffusion Transformer for Talking Video Synthesis
[Code
[Project
Implicit Face Motion Diffusion Model for High-Fidelity Realtime Talking Head Generation
[Project
Audio-Driven Interactive Head Generation in Dyadic Conversations
[Project
Memory-Guided Diffusion for Expressive Talking Video Generation
[Project
[Code
Generative Motion Latent Flow Matching for Audio-driven Talking Portrait
[Project
Highly Dynamic and Realistic Portrait Image Animation with Diffusion Transformer Networks
VQTalker: Towards Multilingual Talking Avatars through Facial Motion Tokenization
Towards Customizable One-Shot Audio-to-Talking Face Generation
IF-MDM: Implicit Face Motion Diffusion Model for High-Fidelity Realtime Talking Head Generation
LatentSync: Audio Conditioned Latent Diffusion Models for Lip Sync
[Code
Diffused Heads: Diffusion Models Beat GANs on Talking-Face Generation
[Project
DiffTalk: Crafting Diffusion Models for Generalized Talking Head Synthesis
[Project
[Code
DAE-Talker: High Fidelity Speech-Driven Talking Face Generation with Diffusion Autoencoder
Emotionally Enhanced Talking Face Generation
[Code
Seeing What You Said: Talking Face Generation Guided by a Lip Reading Expert
[Code
StyleSync: High-Fidelity Generalized and Personalized Lip Sync in Style-based Generator
[Project
[Code
GeneFace++: Generalized and Stable Real-Time Audio-Driven 3D Talking Face Generation
[Project
[Code
MODA: Mapping-Once Audio-driven Portrait Animation with Dual Attentions
VividTalk: One-Shot Audio-Driven Talking Head Generation Based on 3D Hybrid Prior
[Project
[Code
IP_LAP: Identity-Preserving Talking Face Generation with Landmark and Appearance Priors
[Code
HyperLips: Hyper Control Lips with High Resolution Decoder for Talking Face Generation
[Code
Efficient Emotional Adaptation for Audio-Driven Talking-Head Generation
[Project
[Code
SadTalker: Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Talking Head Animation
[Project
[Code
Expressive Talking Head Generation with Granular Audio-Visual Control
Talking Face Generation with Multilingual TTS
[Demo Track
EAMM: One-Shot Emotional Talking Face via Audio-Based Emotion-Aware Motion Model
SPACEx 🚀: Speech-driven Portrait Animation with Controllable Expression
[Project
Masked Lip-Sync Prediction by Audio-Visual Contextual Exploitation in Transformers
Memories are One-to-Many Mapping Alleviators in Talking Face Generation
Pose-Controllable Talking Face Generation by Implicitly Modularized Audio-Visual Representation
[Code
[Project
Imitating Arbitrary Talking Style for Realistic Audio-Driven Talking Face Synthesis
Audio-Driven Emotional Video Portraits
[Code
Talking Head Generation with Audio and Speech Related Facial Action Units
Speech2Talking-Face: Inferring and Driving a Face with Synchronized Audio-Visual Representation
Imitating Arbitrary Talking Style for Realistic Audio-Driven Talking Face Synthesis
Live Speech Portraits: Real-Time Photorealistic Talking-Head Animation
[Code
Audio2head: Audio-driven one-shot talking-head generation with natural head motion
A Lip Sync Expert Is All You Need for Speech to Lip Generation In The Wild
[Code
[Project
Talking-head Generation with Rhythmic Head Motion
[Code
MakeItTalk: Speaker-Aware Talking-Head Animation
[Code
[Project
Neural Voice Puppetry: Audio-driven Facial Reenactment
[Code
[Project
MEAD: A Large-scale Audio-visual Dataset for Emotional Talking-face Generation
[Code
[Project
Realistic Speech-Driven Facial Animation with GANs
Talking Face Generation by Adversarially Disentangled Audio-Visual Representation
[Code
Hierarchical Cross-modal Talking Face Generation with Dynamic Pixel-wise Loss
[Code
Lip Movements Generation at a Glance
[Code
VisemeNet: Audio-Driven Animator-Centric Speech Animation
Synthesizing Obama: Learning Lip Sync From Audio
[Project
You Said That?: Synthesising Talking Faces From Audio
[Code
Audio-Driven Facial Animation by Joint End-to-End Learning of Pose and Emotion
A Deep Learning Approach for Generalized Speech Animation
Lip Reading in the Wild
CVTHead: One-shot Controllable Head Avatar with Vertex-feature Transformer
[Code
3D-Aware Talking-Head Video Motion Transfer
Semantic-Aware Implicit Neural Audio-Driven Video Portrait Generation
HeadNeRF: A Real-time NeRF-based Parametric Head Model
[Code
[Project
I M Avatar: Implicit Morphable Head Avatars from Videos
[Code
Realistic One-shot Mesh-based Head Avatars
FNeVR: Neural Volume Rendering for Face Animation
[Code
3DFaceShop: Explicitly Controllable 3D-Aware Portrait Generation
[Code
[Project
Generative Neural Texture Rasterization for 3D-Aware Head Avatars
[Project
NeRFInvertor: High Fidelity NeRF-GAN Inversion for Single-shot Real Image Animation
Learning Dynamic Facial Radiance Fields for Few-Shot Talking Head Synthesis
[Code
DFA-NeRF: Personalized Talking Head Generation via Disentangled Face Attributes Neural Rendering
NerFACE: Dynamic Neural Radiance Fields for Monocular 4D Facial Avatar Reconstruction
[Code
[Project
AD-NeRF: Audio Driven Neural Radiance Fields for Talking Head Synthesis
[Code
[Code
Disentangled and Controllable Face Image Generation via 3D Imitative-Contrastive Learning
[Code
copy url
source logoblenderartists.org
Joshea
Joshea
copy url
source logolinkedin.com
Learn more
copy url
source logogodotengine.org
copy url
source logogithub.com
Rhubarb Lip Sync
(see below)
Table of Contents
Usage
Speech Recognizer
Mouth Libraries
Cleaning Routine
File Checking Timer
Screenshots
Videos
FileSelectorPreview
Audio Slicing
Calling functions by Code.
Function Usage
Disclaimer
Bug Report
That's because the plugin relies on a Timer instead of communicating with Rhubarb.
Rhubarb Lip Sync's section on Recognizers
below
copy url
source logogithub.com
below
below
below
below
external link
output formats
TSV
XML
JSON
follow this link
Command-line options
follow this link
follow this link
latest release
command-line options
TSV format
XML format
consoleLevel
adheres to SemVer
official SemVer website
Create an issue!
copy url
source logogithub.com
Rhubarb Lip Sync
Blender
https://github.com/scaredyfish/blender-rhubarb-lipsync/issues
http://www.youtube.com/watch?v=azrpByrvw-o
https://github.com/scaredyfish/blender-rhubarb-lipsync/releases
Rhubarb Lip Sync repository
pose library
copy url
source logox.com
copy url
source logogithub.com
below
below
below
below
external link
output formats
TSV
XML
JSON
follow this link
Command-line options
follow this link
follow this link
latest release
command-line options
TSV format
XML format
consoleLevel
adheres to SemVer
official SemVer website
Create an issue!