no code implementations • ECCV 2020 • Bichuan Guo, Jiangtao Wen, Yuxing Han
Light-field cameras capture sub-views from multiple perspectives simultaneously, with possibly reflectance variations that can be used to augment material recognition in remote sensing, autonomous driving, etc.
no code implementations • 13 Mar 2024 • Yuxing Han, Yunan Ding, Chen Ye Gan, Jiangtao Wen
Classifying videos into distinct categories, such as Sport and Music Video, is crucial for multimedia understanding and retrieval, especially when an immense volume of video content is being constantly generated.
no code implementations • 14 Sep 2023 • Yuxing Han, Yunan Ding, Jiangtao Wen, Chen Ye Gan
Classifying videos into distinct categories, such as Sport and Music Video, is crucial for multimedia understanding and retrieval, especially in an age where an immense volume of video content is constantly being generated.
1 code implementation • CVPR 2023 • Yubin Hu, Yuze He, Yanghao Li, Jisheng Li, Yuxing Han, Jiangtao Wen, Yong-Jin Liu
In this paper, we propose an altering resolution framework called AR-Seg for compressed videos to achieve efficient VSS.
no code implementations • 9 Jan 2022 • Xinrong Zhang, Zihou Ren, Xi Li, Shuqi Liu, Yunlong Deng, Yadi Xiao, Yuxing Han, Jiangtao Wen
The global influential factor of the reference to the citing paper is the product of the local influential factor and the total influential factor of the citing paper.
no code implementations • 10 Mar 2021 • Jisheng Li, Yuze He, Yubin Hu, Yuxing Han, Jiangtao Wen
The system utilizes conventional omnidirectional VR camera footage directly without the need for a depth map or segmentation mask, thereby significantly simplifying the overall complexity of the 6-DoF omnidirectional video composition.
no code implementations • 10 Mar 2021 • Jisheng Li, Ziyu Wen, Sihan Li, Yikai Zhao, Bichuan Guo, Jiangtao Wen
Regular omnidirectional video encoding technics use map projection to flatten a scene from a spherical shape into one or several 2D shapes.
no code implementations • 10 Mar 2021 • Jisheng Li, Qi Dai, Jiangtao Wen
Consistent in-focus input imagery is an essential precondition for machine vision systems to perceive the dynamic environment.
no code implementations • ECCV 2020 • Jiangyue Xia, Anyi Rao, Qingqiu Huang, Linning Xu, Jiangtao Wen, Dahua Lin
The task of searching certain people in videos has seen increasing potential in real-world applications, such as video organization and editing.
no code implementations • 7 Jul 2020 • Yanghao Li, Bichuan Guo, Jiangtao Wen, Zhen Xia, Shan Liu, Yuxing Han
Denoisers trained with synthetic data often fail to cope with the diversity of unknown noises, giving way to methods that can adapt to existing noise without knowing its ground truth.
1 code implementation • NeurIPS 2019 • Bichuan Guo, Yuxing Han, Jiangtao Wen
In this paper we propose to use a denoising autoencoder (DAE) prior to simultaneously solve a linear inverse problem and estimate its noise parameter.
2 code implementations • IJCNLP 2019 • Zhihong Shao, Minlie Huang, Jiangtao Wen, Wenfei Xu, Xiaoyan Zhu
Existing neural methods for data-to-text generation are still struggling to produce long and diverse texts: they are insufficient to model input data dynamically during generation, to capture inter-sentence coherence, or to generate diversified expressions.