Talking Head Generation

40 papers with code • 7 benchmarks • 3 datasets

Talking head generation is the task of generating a talking face from a set of images of a person.

( Image credit: Few-Shot Adversarial Learning of Realistic Neural Talking Head Models )

Latest papers with no code

EDTalk: Efficient Disentanglement for Emotional Talking Head Synthesis

no code yet • 2 Apr 2024

Achieving disentangled control over multiple facial motions and accommodating diverse input modalities greatly enhances the application and entertainment of the talking head generation.

EmoVOCA: Speech-Driven Emotional 3D Talking Heads

no code yet • 19 Mar 2024

The domain of 3D talking head generation has witnessed significant progress in recent years.

Style2Talker: High-Resolution Talking Head Generation with Emotion Style and Art Style

no code yet • 11 Mar 2024

Although automatically animating audio-driven talking heads has recently received growing interest, previous efforts have mainly concentrated on achieving lip synchronization with the audio, neglecting two crucial elements for generating expressive videos: emotion style and art style.

AE-NeRF: Audio Enhanced Neural Radiance Field for Few Shot Talking Head Synthesis

no code yet • 18 Dec 2023

Audio-driven talking head synthesis is a promising topic with wide applications in digital human, film making and virtual reality.

VectorTalker: SVG Talking Face Generation with Progressive Vectorisation

no code yet • 18 Dec 2023

To address these, we propose a novel scalable vector graphic reconstruction and animation method, dubbed VectorTalker.

DreamTalk: When Expressive Talking Head Generation Meets Diffusion Probabilistic Models

no code yet • 15 Dec 2023

In this work, we propose a DreamTalk framework to fulfill this gap, which employs meticulous design to unlock the potential of diffusion models in generating expressive talking heads.

PMMTalk: Speech-Driven 3D Facial Animation from Complementary Pseudo Multi-modal Features

no code yet • 5 Dec 2023

The framework entails three modules: PMMTalk encoder, cross-modal alignment module, and PMMTalk decoder.

VividTalk: One-Shot Audio-Driven Talking Head Generation Based on 3D Hybrid Prior

no code yet • 4 Dec 2023

Audio-driven talking head generation has drawn much attention in recent years, and many efforts have been made in lip-sync, expressive facial expressions, natural head pose generation, and high video quality.

LaughTalk: Expressive 3D Talking Head Generation with Laughter

no code yet • 2 Nov 2023

Laughter is a unique expression, essential to affirmative social interactions of humans.

OSM-Net: One-to-Many One-shot Talking Head Generation with Spontaneous Head Motions

no code yet • 28 Sep 2023

Other works construct one-to-one mapping between audio signal and head motion sequences, introducing ambiguity correspondences into the mapping since people can behave differently in head motions when speaking the same content.