Search Results for author: Yabo Zhang

Found 8 papers, 6 papers with code

MC$^2$: Multi-concept Guidance for Customized Multi-concept Generation

1 code implementation8 Apr 2024 Jiaxiu Jiang, Yabo Zhang, Kailai Feng, Xiaohe Wu, WangMeng Zuo

Customized text-to-image generation aims to synthesize instantiations of user-specified concepts and has achieved unprecedented progress in handling individual concept.

Text-to-Image Generation

VideoElevator: Elevating Video Generation Quality with Versatile Text-to-Image Diffusion Models

1 code implementation8 Mar 2024 Yabo Zhang, Yuxiang Wei, Xianhui Lin, Zheng Hui, Peiran Ren, Xuansong Xie, Xiangyang Ji, WangMeng Zuo

Different from conventional T2V sampling (i. e., temporal and spatial modeling), VideoElevator explicitly decomposes each sampling step into temporal motion refining and spatial quality elevating.

Video Generation

Ref-Diff: Zero-shot Referring Image Segmentation with Generative Models

no code implementations31 Aug 2023 Minheng Ni, Yabo Zhang, Kailai Feng, Xiaoming Li, Yiwen Guo, WangMeng Zuo

In this work, we introduce a novel Referring Diffusional segmentor (Ref-Diff) for this task, which leverages the fine-grained multi-modal information from generative models.

Image Segmentation Instance Segmentation +2

VQ-Font: Few-Shot Font Generation with Structure-Aware Enhancement and Quantization

1 code implementation27 Aug 2023 Mingshuai Yao, Yabo Zhang, Xianhui Lin, Xiaoming Li, WangMeng Zuo

In this paper, we propose a VQGAN-based framework (i. e., VQ-Font) to enhance glyph fidelity through token prior refinement and structure-aware enhancement.

Font Generation Quantization

ControlVideo: Training-free Controllable Text-to-Video Generation

1 code implementation22 May 2023 Yabo Zhang, Yuxiang Wei, Dongsheng Jiang, Xiaopeng Zhang, WangMeng Zuo, Qi Tian

Text-driven diffusion models have unlocked unprecedented abilities in image generation, whereas their video counterpart still lags behind due to the excessive training cost of temporal modeling.

Image Generation Text-to-Video Generation +1

ELITE: Encoding Visual Concepts into Textual Embeddings for Customized Text-to-Image Generation

1 code implementation ICCV 2023 Yuxiang Wei, Yabo Zhang, Zhilong Ji, Jinfeng Bai, Lei Zhang, WangMeng Zuo

In addition to the unprecedented ability in imaginary creation, large text-to-image models are expected to take customized concepts in image generation.

Text-to-Image Generation

Towards Diverse and Faithful One-shot Adaption of Generative Adversarial Networks

1 code implementation18 Jul 2022 Yabo Zhang, Mingshuai Yao, Yuxiang Wei, Zhilong Ji, Jinfeng Bai, WangMeng Zuo

In this paper, we present a novel one-shot generative domain adaption method, i. e., DiFa, for diverse generation and faithful adaptation.

Domain Adaptation

Cannot find the paper you are looking for? You can Submit a new open access paper.