Talking Face Generation

37 papers with code • 2 benchmarks • 6 datasets

Talking face generation aims to synthesize a sequence of face images that correspond to given speech semantics

( Image credit: Talking Face Generation by Adversarially Disentangled Audio-Visual Representation )

Latest papers with no code

Superior and Pragmatic Talking Face Generation with Teacher-Student Framework

no code yet • 26 Mar 2024

Talking face generation technology creates talking videos from arbitrary appearance and motion signal, with the "arbitrary" offering ease of use but also introducing challenges in practical applications.

FlowVQTalker: High-Quality Emotional Talking Face Generation through Normalizing Flow and Quantization

no code yet • 11 Mar 2024

Specifically, we develop a flow-based coefficient generator that encodes the dynamics of facial emotion into a multi-emotion-class latent space represented as a mixture distribution.

Style2Talker: High-Resolution Talking Head Generation with Emotion Style and Art Style

no code yet • 11 Mar 2024

Although automatically animating audio-driven talking heads has recently received growing interest, previous efforts have mainly concentrated on achieving lip synchronization with the audio, neglecting two crucial elements for generating expressive videos: emotion style and art style.

G4G:A Generic Framework for High Fidelity Talking Face Generation with Fine-grained Intra-modal Alignment

no code yet • 28 Feb 2024

Despite numerous completed studies, achieving high fidelity talking face generation with highly synchronized lip movements corresponding to arbitrary audio remains a significant challenge in the field.

AVI-Talking: Learning Audio-Visual Instructions for Expressive 3D Talking Face Generation

no code yet • 25 Feb 2024

In this paper, we propose AVI-Talking, an Audio-Visual Instruction system for expressive Talking face generation.

EmoSpeaker: One-shot Fine-grained Emotion-Controlled Talking Face Generation

no code yet • 2 Feb 2024

Additionally, we propose an emotion intensity control method using a fine-grained emotion matrix.

EmoTalker: Emotionally Editable Talking Face Generation via Diffusion Model

no code yet • 16 Jan 2024

In recent years, the field of talking faces generation has attracted considerable attention, with certain methods adept at generating virtual faces that convincingly imitate human expressions.

DREAM-Talk: Diffusion-based Realistic Emotional Audio-driven Method for Single Image Talking Face Generation

no code yet • 21 Dec 2023

The generation of emotional talking faces from a single portrait image remains a significant challenge.

VectorTalker: SVG Talking Face Generation with Progressive Vectorisation

no code yet • 18 Dec 2023

To address these, we propose a novel scalable vector graphic reconstruction and animation method, dubbed VectorTalker.

GSmoothFace: Generalized Smooth Talking Face Generation via Fine Grained 3D Face Guidance

no code yet • 12 Dec 2023

Our proposed GSmoothFace model mainly consists of the Audio to Expression Prediction (A2EP) module and the Target Adaptive Face Translation (TAFT) module.