Talking Face Generation
37 papers with code • 2 benchmarks • 6 datasets
Talking face generation aims to synthesize a sequence of face images that correspond to given speech semantics
( Image credit: Talking Face Generation by Adversarially Disentangled Audio-Visual Representation )
Latest papers with no code
Superior and Pragmatic Talking Face Generation with Teacher-Student Framework
Talking face generation technology creates talking videos from arbitrary appearance and motion signal, with the "arbitrary" offering ease of use but also introducing challenges in practical applications.
FlowVQTalker: High-Quality Emotional Talking Face Generation through Normalizing Flow and Quantization
Specifically, we develop a flow-based coefficient generator that encodes the dynamics of facial emotion into a multi-emotion-class latent space represented as a mixture distribution.
Style2Talker: High-Resolution Talking Head Generation with Emotion Style and Art Style
Although automatically animating audio-driven talking heads has recently received growing interest, previous efforts have mainly concentrated on achieving lip synchronization with the audio, neglecting two crucial elements for generating expressive videos: emotion style and art style.
G4G:A Generic Framework for High Fidelity Talking Face Generation with Fine-grained Intra-modal Alignment
Despite numerous completed studies, achieving high fidelity talking face generation with highly synchronized lip movements corresponding to arbitrary audio remains a significant challenge in the field.
AVI-Talking: Learning Audio-Visual Instructions for Expressive 3D Talking Face Generation
In this paper, we propose AVI-Talking, an Audio-Visual Instruction system for expressive Talking face generation.
EmoSpeaker: One-shot Fine-grained Emotion-Controlled Talking Face Generation
Additionally, we propose an emotion intensity control method using a fine-grained emotion matrix.
EmoTalker: Emotionally Editable Talking Face Generation via Diffusion Model
In recent years, the field of talking faces generation has attracted considerable attention, with certain methods adept at generating virtual faces that convincingly imitate human expressions.
DREAM-Talk: Diffusion-based Realistic Emotional Audio-driven Method for Single Image Talking Face Generation
The generation of emotional talking faces from a single portrait image remains a significant challenge.
VectorTalker: SVG Talking Face Generation with Progressive Vectorisation
To address these, we propose a novel scalable vector graphic reconstruction and animation method, dubbed VectorTalker.
GSmoothFace: Generalized Smooth Talking Face Generation via Fine Grained 3D Face Guidance
Our proposed GSmoothFace model mainly consists of the Audio to Expression Prediction (A2EP) module and the Target Adaptive Face Translation (TAFT) module.