no code implementations • 2 Jan 2024 • Zhichao Yin, Binyuan Hui, Min Yang, Fei Huang, Yongbin Li
Recently, substantial advancements in pre-trained vision-language models have greatly enhanced the capabilities of multi-modal dialog systems.
1 code implementation • 24 May 2023 • Yunshui Li, Binyuan Hui, Zhichao Yin, Min Yang, Fei Huang, Yongbin Li
It utilizes a combination of several fundamental experts to accommodate multiple dialogue-related tasks and can be pre-trained using limited dialogue and extensive non-dialogue multi-modal data.
Ranked #1 on Response Generation on SIMMC2.0
no code implementations • 16 Nov 2022 • Jun Zhou, Zhichao Yin, Pengpeng Yue
This paper proposes a brand-new measure of energy efficiency at household level and explores how it is affected by access to credit.
no code implementations • 23 Jan 2022 • Pengpeng Yue, Aslihan Gizem Korkmaz, Zhichao Yin, Haigang Zhou
However, the easier access to credit markets also increases the risk of households falling into a debt trap.
2 code implementations • CVPR 2019 • Zhichao Yin, Trevor Darrell, Fisher Yu
Explicit representations of the global match distributions of pixel-wise correspondences between pairs of images are desirable for uncertainty estimation and downstream applications.
Ranked #13 on Optical Flow Estimation on KITTI 2015 (train)
no code implementations • 1 Aug 2018 • Xinge Zhu, Zhichao Yin, Jianping Shi, Hongsheng Li, Dahua Lin
Due to the large gap and severe deformation between the frontal view and bird view, generating a bird view image from a single frontal view is challenging.
3 code implementations • CVPR 2018 • Zhichao Yin, Jianping Shi
We propose GeoNet, a jointly unsupervised learning framework for monocular depth, optical flow and ego-motion estimation from videos.
Ranked #1 on Pose Estimation on KITTI 2015