Scene Generation
57 papers with code • 3 benchmarks • 6 datasets
Libraries
Use these libraries to find Scene Generation models and implementationsLatest papers
SemCity: Semantic Scene Generation with Triplane Diffusion
In this paper, we concentrate on generating a real-outdoor scene through learning a diffusion model on a real-world outdoor dataset.
WoVoGen: World Volume-aware Diffusion for Controllable Multi-camera Driving Scene Generation
Generating multi-camera street-view videos is critical for augmenting autonomous driving datasets, addressing the urgent demand for extensive and varied data.
Pyramid Diffusion for Fine 3D Large Scene Generation
Directly transferring the 2D techniques to 3D scene generation is challenging due to significant resolution reduction and the scarcity of comprehensive real-world 3D scene datasets.
High-fidelity Person-centric Subject-to-Image Synthesis
Specifically, we first develop two specialized pre-trained diffusion models, i. e., Text-driven Diffusion Model (TDM) and Subject-augmented Diffusion Model (SDM), for scene and person generation, respectively.
LLM Blueprint: Enabling Text-to-Image Generation with Complex and Detailed Prompts
Diffusion-based generative models have significantly advanced text-to-image generation but encounter challenges when processing lengthy and intricate text prompts describing complex scenes with multiple objects.
RoomDesigner: Encoding Anchor-latents for Style-consistent and Shape-compatible Indoor Scene Generation
Indoor scene generation aims at creating shape-compatible, style-consistent furniture arrangements within a spatially reasonable layout.
On the Generation of a Synthetic Event-Based Vision Dataset for Navigation and Landing
We anticipate that novel event-based vision datasets can be generated using this pipeline to support various spacecraft pose reconstruction problems given events as input, and we hope that the proposed methodology would attract the attention of researchers working at the intersection of neuromorphic vision and guidance navigation and control.
Language Conditioned Traffic Generation
In this work, we turn to language as a source of supervision for dynamic traffic scene generation.
DiffInDScene: Diffusion-based High-Quality 3D Indoor Scene Generation
In addition to the scene generation, the final part of DiffInDScene can be used as a post-processing module to refine the 3D reconstruction results from multi-view stereo.
CommonScenes: Generating Commonsense 3D Indoor Scenes with Scene Graph Diffusion
The generated scenes can be manipulated by editing the input scene graph and sampling the noise in the diffusion model.