Search Results for author: Jen-Hao Rick Chang

Found 17 papers, 2 papers with code

Probabilistic Speech-Driven 3D Facial Motion Synthesis: New Benchmarks, Methods, and Applications

no code implementations30 Nov 2023 Karren D. Yang, Anurag Ranjan, Jen-Hao Rick Chang, Raviteja Vemulapalli, Oncel Tuzel

While these models can achieve high-quality lip articulation for speakers in the training set, they are unable to capture the full and diverse distribution of 3D facial motions that accompany speech in the real world.

Motion Synthesis

HUGS: Human Gaussian Splats

1 code implementation29 Nov 2023 Muhammed Kocabas, Jen-Hao Rick Chang, James Gabriel, Oncel Tuzel, Anurag Ranjan

We achieve state-of-the-art rendering quality with a rendering speed of 60 FPS while being ~100x faster to train over previous work.

Neural Rendering Novel View Synthesis

Novel-View Acoustic Synthesis from 3D Reconstructed Rooms

1 code implementation23 Oct 2023 Byeongjoo Ahn, Karren Yang, Brian Hamilton, Jonathan Sheaffer, Anurag Ranjan, Miguel Sarabia, Oncel Tuzel, Jen-Hao Rick Chang

Given audio recordings from 2-4 microphones and the 3D geometry and material of a scene containing multiple unknown sound sources, we estimate the sound anywhere in the scene.

Efficient-3DiM: Learning a Generalizable Single-image Novel-view Synthesizer in One Day

no code implementations4 Oct 2023 Yifan Jiang, Hao Tang, Jen-Hao Rick Chang, Liangchen Song, Zhangyang Wang, Liangliang Cao

Although the fidelity and generalizability are greatly improved, training such a powerful diffusion model requires a vast volume of training data and model parameters, resulting in a notoriously long time and high computational costs.

Image Generation Novel View Synthesis

Pointersect: Neural Rendering with Cloud-Ray Intersection

no code implementations CVPR 2023 Jen-Hao Rick Chang, Wei-Yu Chen, Anurag Ranjan, Kwang Moo Yi, Oncel Tuzel

Specifically, we train a set transformer that, given a small number of local neighbor points along a light ray, provides the intersection point, the surface normal, and the material blending weights, which are used to render the outcome of this light ray.

Inverse Rendering Neural Rendering +2

FaceLit: Neural 3D Relightable Faces

no code implementations CVPR 2023 Anurag Ranjan, Kwang Moo Yi, Jen-Hao Rick Chang, Oncel Tuzel

We propose a generative framework, FaceLit, capable of generating a 3D face that can be rendered at various user-defined lighting conditions and views, learned purely from 2D images in-the-wild without any manual annotation.

Text is All You Need: Personalizing ASR Models using Controllable Speech Synthesis

no code implementations27 Mar 2023 Karren Yang, Ting-yao Hu, Jen-Hao Rick Chang, Hema Swetha Koppula, Oncel Tuzel

Here, we ask two fundamental questions about this strategy: when is synthetic data effective for personalization, and why is it effective in those cases?

Automatic Speech Recognition Automatic Speech Recognition (ASR) +3

Data Incubation -- Synthesizing Missing Data for Handwriting Recognition

no code implementations13 Oct 2021 Jen-Hao Rick Chang, Martin Bresler, Youssouf Chherawala, Adrien Delaye, Thomas Deselaers, Ryan Dixon, Oncel Tuzel

We use the framework to optimize data synthesis and demonstrate significant improvement on handwriting recognition over a model trained on real data only.

Handwriting Recognition

Token Pooling in Vision Transformers

no code implementations8 Oct 2021 Dmitrii Marin, Jen-Hao Rick Chang, Anurag Ranjan, Anish Prabhu, Mohammad Rastegari, Oncel Tuzel

Token Pooling is a simple and effective operator that can benefit many architectures.

Style Equalization: Unsupervised Learning of Controllable Generative Sequence Models

no code implementations6 Oct 2021 Jen-Hao Rick Chang, Ashish Shrivastava, Hema Swetha Koppula, Xiaoshuai Zhang, Oncel Tuzel

However, under an unsupervised-style setting, typical training algorithms for controllable sequence generative models suffer from the training-inference mismatch, where the same sample is used as content and style input during training but unpaired samples are given during inference.

Towards Occlusion-Aware Multifocal Displays

no code implementations2 May 2020 Jen-Hao Rick Chang, Anat Levin, B. V. K. Vijaya Kumar, Aswin C. Sankaranarayanan

Multifocal displays, one of the classic approaches to satisfy the accommodation cue, place virtual content at multiple focal planes, each at a di erent depth.

Towards Multifocal Displays with Dense Focal Stacks

no code implementations27 May 2018 Jen-Hao Rick Chang, B. V. K. Vijaya Kumar, Aswin C. Sankaranarayanan

We present a virtual reality display that is capable of generating a dense collection of depth/focal planes.

Propagated Image Filtering

no code implementations CVPR 2015 Jen-Hao Rick Chang, Yu-Chiang Frank Wang

In this paper, we propose the propagation filter as a novel image filtering operator, with the goal of smoothing over neighboring image pixels while preserving image context like edges or textural regions.

Image Denoising

Cannot find the paper you are looking for? You can Submit a new open access paper.