Image Generation
1990 papers with code • 85 benchmarks • 67 datasets
Image Generation (synthesis) is the task of generating new images from an existing dataset.
- Unconditional generation refers to generating samples unconditionally from the dataset, i.e. $p(y)$
- Conditional image generation (subtask) refers to generating samples conditionally from the dataset, based on a label, i.e. $p(y|x)$.
In this section, you can find state-of-the-art leaderboards for unconditional generation. For conditional generation, and other types of image generations, refer to the subtasks.
( Image credit: StyleGAN )
Libraries
Use these libraries to find Image Generation models and implementationsDatasets
Subtasks
- Image-to-Image Translation
- Image Inpainting
- Text-to-Image Generation
- Conditional Image Generation
- Conditional Image Generation
- Face Generation
- 3D Generation
- Image Harmonization
- Pose Transfer
- 3D-Aware Image Synthesis
- Facial Inpainting
- Layout-to-Image Generation
- ROI-based image generation
- Image Generation from Scene Graphs
- Pose-Guided Image Generation
- User Constrained Thumbnail Generation
- Handwritten Word Generation
- Chinese Landscape Painting Generation
- person reposing
- Infinite Image Generation
- Multi class one-shot image synthesis
- Single class few-shot image synthesis
Latest papers
ANCHOR: LLM-driven News Subject Conditioning for Text-to-Image Synthesis
With Large Language Models (LLM) achieving success in language and commonsense reasoning tasks, we explore the ability of different LLMs to identify and understand key subjects from abstractive captions.
Semantic Approach to Quantifying the Consistency of Diffusion Model Image Generation
In this study, we identify the need for an interpretable, quantitative score of the repeatability, or consistency, of image generation in diffusion models.
CAT: Contrastive Adapter Training for Personalized Image Generation
Finally, we mention the possibility of CAT in the aspects of multi-concept adapter and optimization.
Taming Stable Diffusion for Text to 360° Panorama Image Generation
Generative models, e. g., Stable Diffusion, have enabled the creation of photorealistic images from text prompts.
Latent Guard: a Safety Framework for Text-to-image Generation
Hence, we propose Latent Guard, a framework designed to improve safety measures in text-to-image generation.
Model-based Cleaning of the QUILT-1M Pathology Dataset for Text-Conditional Image Synthesis
The QUILT-1M dataset is the first openly available dataset containing images harvested from various online sources.
A Gauss-Newton Approach for Min-Max Optimization in Generative Adversarial Networks
It modifies the Gauss-Newton method to approximate the min-max Hessian and uses the Sherman-Morrison inversion formula to calculate the inverse.
Deep Generative Data Assimilation in Multimodal Setting
To our knowledge, our work is the first to apply deep generative framework for multimodal data assimilation using real-world datasets; an important step for building robust computational simulators, including the next-generation Earth system models.
StoryImager: A Unified and Efficient Framework for Coherent Story Visualization and Completion
3) The story visualization and continuation models are trained and inferred independently, which is not user-friendly.
Hyperparameter-Free Medical Image Synthesis for Sharing Data and Improving Site-Specific Segmentation
Sharing synthetic medical images is a promising alternative to sharing real images that can improve patient privacy and data security.