Search Results for author: Ligong Han

Found 24 papers, 14 papers with code

Score-Guided Diffusion for 3D Human Recovery

1 code implementation14 Mar 2024 Anastasis Stathopoulos, Ligong Han, Dimitris Metaxas

We present Score-Guided Human Mesh Recovery (ScoreHMR), an approach for solving inverse problems for 3D human pose and shape reconstruction.

Denoising Human Mesh Recovery

DMCVR: Morphology-Guided Diffusion Model for 3D Cardiac Volume Reconstruction

1 code implementation18 Aug 2023 Xiaoxiao He, Chaowei Tan, Ligong Han, Bo Liu, Leon Axel, Kang Li, Dimitris N. Metaxas

However, current cardiac MRI-based reconstruction technology used in clinical settings is 2D with limited through-plane resolution, resulting in low-quality reconstructed cardiac volumes.

3D Reconstruction

Improving Tuning-Free Real Image Editing with Proximal Guidance

1 code implementation8 Jun 2023 Ligong Han, Song Wen, Qi Chen, Zhixing Zhang, Kunpeng Song, Mengwei Ren, Ruijiang Gao, Anastasis Stathopoulos, Xiaoxiao He, Yuxiao Chen, Di Liu, Qilong Zhangli, Jindong Jiang, Zhaoyang Xia, Akash Srivastava, Dimitris Metaxas

Null-text inversion (NTI) optimizes null embeddings to align the reconstruction and inversion trajectories with larger CFG scales, enabling real image editing with cross-attention control.

Learning Articulated Shape with Keypoint Pseudo-labels from Web Images

no code implementations CVPR 2023 Anastasis Stathopoulos, Georgios Pavlakos, Ligong Han, Dimitris Metaxas

It is based on two key insights: (1) 2D keypoint estimation networks trained on as few as 50-150 images of a given object category generalize well and generate reliable pseudo-labels; (2) a data selection mechanism can automatically create a "curated" subset of the unlabeled web images that can be used for training -- we evaluate four data selection methods.

3D Reconstruction Keypoint Estimation +1

SVDiff: Compact Parameter Space for Diffusion Fine-Tuning

1 code implementation ICCV 2023 Ligong Han, Yinxiao Li, Han Zhang, Peyman Milanfar, Dimitris Metaxas, Feng Yang

Diffusion models have achieved remarkable success in text-to-image generation, enabling the creation of high-quality images from text prompts or other modalities.

Data Augmentation Efficient Diffusion Personalization +1

Learning Complementary Policies for Human-AI Teams

no code implementations6 Feb 2023 Ruijiang Gao, Maytal Saar-Tsechansky, Maria De-Arteaga, Ligong Han, Wei Sun, Min Kyung Lee, Matthew Lease

We then extend our approach to leverage opportunities and mitigate risks that arise in important contexts in practice: 1) when a team is composed of multiple humans with differential and potentially complementary abilities, 2) when the observational data includes consistent deterministic actions, and 3) when the covariate distribution of future decisions differ from that in the historical data.

SINE: SINgle Image Editing with Text-to-Image Diffusion Models

1 code implementation CVPR 2023 Zhixing Zhang, Ligong Han, Arnab Ghosh, Dimitris Metaxas, Jian Ren

We propose a novel model-based guidance built upon the classifier-free guidance so that the knowledge from the model trained on a single image can be distilled into the pre-trained diffusion model, enabling content creation even with one given image.

Image Generation

Diffusion Guided Domain Adaptation of Image Generators

no code implementations8 Dec 2022 Kunpeng Song, Ligong Han, Bingchen Liu, Dimitris Metaxas, Ahmed Elgammal

Can a text-to-image diffusion model be used as a training objective for adapting a GAN generator to another domain?

Domain Adaptation

On the Importance of Calibration in Semi-supervised Learning

no code implementations10 Oct 2022 Charlotte Loh, Rumen Dangovski, Shivchander Sudalairaj, Seungwook Han, Ligong Han, Leonid Karlinsky, Marin Soljacic, Akash Srivastava

State-of-the-art (SOTA) semi-supervised learning (SSL) methods have been highly successful in leveraging a mix of labeled and unlabeled data by combining techniques of consistency regularization and pseudo-labeling.

Hierarchically Self-Supervised Transformer for Human Skeleton Representation Learning

1 code implementation20 Jul 2022 Yuxiao Chen, Long Zhao, Jianbo Yuan, Yu Tian, Zhaoyang Xia, Shijie Geng, Ligong Han, Dimitris N. Metaxas

Despite the success of fully-supervised human skeleton sequence modeling, utilizing self-supervised pre-training for skeleton sequence representation learning has been an active field because acquiring task-specific skeleton annotations at large scales is difficult.

Action Detection Action Recognition +3

Show Me What and Tell Me How: Video Synthesis via Multimodal Conditioning

1 code implementation CVPR 2022 Ligong Han, Jian Ren, Hsin-Ying Lee, Francesco Barbieri, Kyle Olszewski, Shervin Minaee, Dimitris Metaxas, Sergey Tulyakov

In addition, our model can extract visual information as suggested by the text prompt, e. g., "an object in image one is moving northeast", and generate corresponding videos.

Self-Learning Text Augmentation +1

Enhancing Counterfactual Classification via Self-Training

1 code implementation8 Dec 2021 Ruijiang Gao, Max Biggs, Wei Sun, Ligong Han

We approach this task as a domain adaptation problem and propose a self-training algorithm which imputes outcomes with categorical values for finite unseen actions in the observational data to simulate a randomized trial through pseudolabeling, which we refer to as Counterfactual Self-Training (CST).

Classification counterfactual +2

AE-StyleGAN: Improved Training of Style-Based Auto-Encoders

1 code implementation17 Oct 2021 Ligong Han, Sri Harsha Musunuri, Martin Renqiang Min, Ruijiang Gao, Yu Tian, Dimitris Metaxas

StyleGANs have shown impressive results on data generation and manipulation in recent years, thanks to its disentangled style latent space.

Dual Projection Generative Adversarial Networks for Conditional Image Generation

1 code implementation ICCV 2021 Ligong Han, Martin Renqiang Min, Anastasis Stathopoulos, Yu Tian, Ruijiang Gao, Asim Kadav, Dimitris Metaxas

We then propose an improved cGAN model with Auxiliary Classification that directly aligns the fake and real conditionals $P(\text{class}|\text{image})$ by minimizing their $f$-divergence.

Conditional Image Generation

Disentangled Recurrent Wasserstein Autoencoder

no code implementations ICLR 2021 Jun Han, Martin Renqiang Min, Ligong Han, Li Erran Li, Xuan Zhang

Learning disentangled representations leads to interpretable models and facilitates data generation with style transfer, which has been extensively studied on static data such as images in an unsupervised learning framework.

Disentanglement Style Transfer +1

Counterfactual Self-Training

no code implementations1 Jan 2021 Ruijiang Gao, Max Biggs, Wei Sun, Ligong Han

We approach this task as a domain adaptation problem and propose a self-training algorithm which imputes outcomes for the unseen actions in the observational data to simulate a randomized trial.

counterfactual Domain Adaptation +1

Unbiased Auxiliary Classifier GANs with MINE

1 code implementation13 Jun 2020 Ligong Han, Anastasis Stathopoulos, Tao Xue, Dimitris Metaxas

To remedy this, Twin Auxiliary Classifier GAN (TAC-GAN) introduces a twin classifier to the min-max game.

Learning Generative Models of Tissue Organization with Supervised GANs

1 code implementation31 Mar 2020 Ligong Han, Robert F. Murphy, Deva Ramanan

A key step in understanding the spatial organization of cells and tissues is the ability to construct generative models that accurately reflect that organization.

Image Generation

Robust Conditional GAN from Uncertainty-Aware Pairwise Comparisons

1 code implementation21 Nov 2019 Ligong Han, Ruijiang Gao, Mun Kim, Xin Tao, Bo Liu, Dimitris Metaxas

Conditional generative adversarial networks have shown exceptional generation performance over the past few years.

Attribute Generative Adversarial Network

Unsupervised Domain Adaptation via Calibrating Uncertainties

1 code implementation25 Jul 2019 Ligong Han, Yang Zou, Ruijiang Gao, Lezi Wang, Dimitris Metaxas

Unsupervised domain adaptation (UDA) aims at inferring class labels for unlabeled target domain given a related labeled source dataset.

Unsupervised Domain Adaptation

Deep Contextual Recurrent Residual Networks for Scene Labeling

no code implementations12 Apr 2017 T. Hoang Ngan Le, Chi Nhan Duong, Ligong Han, Khoa Luu, Marios Savvides, Dipan Pal

Designed as extremely deep architectures, deep residual networks which provide a rich visual representation and offer robust convergence behaviors have recently achieved exceptional performance in numerous computer vision problems.

Representation Learning Scene Labeling

Cannot find the paper you are looking for? You can Submit a new open access paper.