Search Results for author: Lizhen Wang

Found 15 papers, 9 papers with code

GMTalker: Gaussian Mixture based Emotional talking video Portraits

no code implementations12 Dec 2023 Yibo Xia, Lizhen Wang, Xiang Deng, Xiaoyan Luo, Yebin Liu

Specifically, we propose a Gaussian Mixture based Expression Generator (GMEG) which can construct a continuous and multi-modal latent space, achieving more flexible emotion manipulation.

MonoGaussianAvatar: Monocular Gaussian Point-based Head Avatar

no code implementations7 Dec 2023 Yufan Chen, Lizhen Wang, Qijing Li, Hongjiang Xiao, Shengping Zhang, Hongxun Yao, Yebin Liu

In response to these challenges, we propose MonoGaussianAvatar (Monocular Gaussian Point-based Head Avatar), a novel approach that harnesses 3D Gaussian point representation coupled with a Gaussian deformation field to learn explicit head avatars from monocular portrait videos.

Gaussian Head Avatar: Ultra High-fidelity Head Avatar via Dynamic Gaussians

1 code implementation5 Dec 2023 Yuelang Xu, Benwang Chen, Zhe Li, Hongwen Zhang, Lizhen Wang, Zerong Zheng, Yebin Liu

Creating high-fidelity 3D head avatars has always been a research hotspot, but there remains a great challenge under lightweight sparse view setups.

2k

InvertAvatar: Incremental GAN Inversion for Generalized Head Avatars

no code implementations3 Dec 2023 Xiaochen Zhao, Jingxiang Sun, Lizhen Wang, Yebin Liu

While high fidelity and efficiency are central to the creation of digital head avatars, recent methods relying on 2D or 3D generative models often experience limitations such as shape distortion, expression inaccuracy, and identity flickering.

Image-to-Image Translation

Animatable Gaussians: Learning Pose-dependent Gaussian Maps for High-fidelity Human Avatar Modeling

1 code implementation27 Nov 2023 Zhe Li, Zerong Zheng, Lizhen Wang, Yebin Liu

Overall, our method can create lifelike avatars with dynamic, realistic and generalized appearances.

DreamCraft3D: Hierarchical 3D Generation with Bootstrapped Diffusion Prior

1 code implementation25 Oct 2023 Jingxiang Sun, Bo Zhang, Ruizhi Shao, Lizhen Wang, Wen Liu, Zhenda Xie, Yebin Liu

The score distillation from this 3D-aware diffusion prior provides view-consistent guidance for the scene.

3D Generation

HAvatar: High-fidelity Head Avatar via Facial Model Conditioned Neural Radiance Field

no code implementations29 Sep 2023 Xiaochen Zhao, Lizhen Wang, Jingxiang Sun, Hongwen Zhang, Jinli Suo, Yebin Liu

The problem of modeling an animatable 3D human head avatar under light-weight setups is of significant importance but has not been well solved.

Image-to-Image Translation

LatentAvatar: Learning Latent Expression Code for Expressive Neural Head Avatar

no code implementations2 May 2023 Yuelang Xu, Hongwen Zhang, Lizhen Wang, Xiaochen Zhao, Han Huang, GuoJun Qi, Yebin Liu

Existing approaches to animatable NeRF-based head avatars are either built upon face templates or use the expression coefficients of templates as the driving signal.

StyleAvatar: Real-time Photo-realistic Portrait Avatar from a Single Video

1 code implementation1 May 2023 Lizhen Wang, Xiaochen Zhao, Jingxiang Sun, Yuxiang Zhang, Hongwen Zhang, Tao Yu, Yebin Liu

Results and experiments demonstrate the superiority of our method in terms of image quality, full portrait video generation, and real-time re-animation compared to existing facial reenactment methods.

Face Reenactment Translation +1

AvatarMAV: Fast 3D Head Avatar Reconstruction Using Motion-Aware Neural Voxels

no code implementations23 Nov 2022 Yuelang Xu, Lizhen Wang, Xiaochen Zhao, Hongwen Zhang, Yebin Liu

AvatarMAV is the first to model both the canonical appearance and the decoupled expression motion by neural voxels for head avatar.

Next3D: Generative Neural Texture Rasterization for 3D-Aware Head Avatars

2 code implementations CVPR 2023 Jingxiang Sun, Xuan Wang, Lizhen Wang, Xiaoyu Li, Yong Zhang, Hongwen Zhang, Yebin Liu

We propose a novel 3D GAN framework for unsupervised learning of generative, high-quality and 3D-consistent facial avatars from unstructured 2D images.

Face Model

IDE-3D: Interactive Disentangled Editing for High-Resolution 3D-aware Portrait Synthesis

1 code implementation31 May 2022 Jingxiang Sun, Xuan Wang, Yichun Shi, Lizhen Wang, Jue Wang, Yebin Liu

Existing 3D-aware facial generation methods face a dilemma in quality versus editability: they either generate editable results in low resolution or high-quality ones with no editing flexibility.

3D-Aware Image Synthesis

FaceVerse: a Fine-grained and Detail-controllable 3D Face Morphable Model from a Hybrid Dataset

1 code implementation CVPR 2022 Lizhen Wang, ZhiYuan Chen, Tao Yu, Chenguang Ma, Liang Li, Yebin Liu

In the coarse module, we generate a base parametric model from large-scale RGB-D images, which is able to predict accurate rough 3D face models in different genders, ages, etc.

2k 3D Face Reconstruction +1

NormalGAN: Learning Detailed 3D Human from a Single RGB-D Image

1 code implementation ECCV 2020 Lizhen Wang, Xiaochen Zhao, Tao Yu, Songtao Wang, Yebin Liu

We propose NormalGAN, a fast adversarial learning-based method to reconstruct the complete and detailed 3D human from a single RGB-D image.

3D Human Reconstruction Denoising

DDRNet: Depth Map Denoising and Refinement for Consumer Depth Cameras Using Cascaded CNNs

3 code implementations ECCV 2018 Shi Yan, Chenglei Wu, Lizhen Wang, Feng Xu, Liang An, Kaiwen Guo, Yebin Liu

Consumer depth sensors are more and more popular and come to our daily lives marked by its recent integration in the latest Iphone X.

Denoising

Cannot find the paper you are looking for? You can Submit a new open access paper.