no code implementations • 12 Apr 2024 • Zeyu Yang, Peikun Guo, Khadija Zanna, Akane Sano
Diffusion models have emerged as a robust framework for various generative tasks, such as image and audio synthesis, and have also demonstrated a remarkable ability to generate mixed-type tabular data comprising both continuous and discrete variables.
1 code implementation • 2 Apr 2024 • Zeyu Yang, Zijie Pan, Chun Gu, Li Zhang
Recent advancements in 3D generation are predominantly propelled by improvements in 3D-aware image diffusion models which are pretrained on Internet-scale image data and fine-tuned on massive 3D data, offering the capability of producing highly consistent multi-view images.
no code implementations • 16 Jan 2024 • Zijie Pan, Zeyu Yang, Xiatian Zhu, Li Zhang
Generating dynamic 3D object from a single-view video is challenging due to the lack of 4D labeled data.
1 code implementation • 5 Dec 2023 • Jiachen Lu, Ze Huang, Zeyu Yang, Jiahui Zhang, Li Zhang
Generating multi-camera street-view videos is critical for augmenting autonomous driving datasets, addressing the urgent demand for extensive and varied data.
1 code implementation • 16 Oct 2023 • Zeyu Yang, Hongye Yang, Zijie Pan, Li Zhang
Reconstructing dynamic 3D scenes from 2D images and generating diverse views over time is challenging due to scene complexity and temporal dynamics.
2 code implementations • 23 Aug 2022 • Zeyu Yang, Jiaqi Chen, Zhenwei Miao, Wei Li, Xiatian Zhu, Li Zhang
Existing top-performance 3D object detectors typically rely on the multi-modal fusion strategy.