no code implementations • 16 Mar 2024 • Deyi Ji, Siqi Gao, Lanyun Zhu, Yiru Zhao, Peng Xu, Hongtao Lu, Feng Zhao
In this paper, we address the challenge of multi-object tracking (MOT) in moving Unmanned Aerial Vehicle (UAV) scenarios, where irregular flight trajectories, such as hovering, turning left/right, and moving up/down, lead to significantly greater complexity compared to fixed-camera MOT.
no code implementations • 28 Feb 2024 • Lanyun Zhu, Deyi Ji, Tianrun Chen, Peng Xu, Jieping Ye, Jun Liu
Despite achieving rapid developments and with widespread applications, Large Vision-Language Models (LVLMs) confront a serious challenge of being prone to generating hallucinations.
no code implementations • 8 Feb 2024 • Ying Zang, Chenglong Fu, Runlong Cao, Didi Zhu, Min Zhang, WenJun Hu, Lanyun Zhu, Tianrun Chen
This pioneering work lays the groundwork for future research in semi-supervised learning for referring expression segmentation.
no code implementations • 28 Nov 2023 • Lanyun Zhu, Tianrun Chen, Deyi Ji, Jieping Ye, Jun Liu
This paper proposes LLaFS, the first attempt to leverage large language models (LLMs) in few-shot segmentation.
no code implementations • 22 Sep 2023 • Tianrun Chen, Chenglong Fu, Ying Zang, Lanyun Zhu, Jia Zhang, Papa Mao, Lingyun Sun
In this work, we introduce a novel end-to-end approach, Deep3DSketch+, which performs 3D modeling using only a single free-hand sketch without inputting multiple sketches or view information.
no code implementations • ICCV 2023 • Lanyun Zhu, Tianrun Chen, Jianxiong Yin, Simon See, Jun Liu
We innovatively utilize Gabor filters as a powerful extractor to exploit texture features, motivated by the capability of Gabor filters in effectively capturing multi-frequency features and detailed local information.
no code implementations • 26 Apr 2023 • Yan Wang, Jian Cheng, Yixin Chen, Shuai Shao, Lanyun Zhu, Zhenzhou Wu, Tao Liu, Haogang Zhu
In FVP, the visual prompt is parameterized using only a small amount of low-frequency learnable parameters in the input frequency space, and is learned by minimizing the segmentation loss between the predicted segmentation of the prompted target image and reliable pseudo segmentation label of the target image under the frozen model.
1 code implementation • 18 Apr 2023 • Tianrun Chen, Lanyun Zhu, Chaotao Ding, Runlong Cao, Yan Wang, Zejian Li, Lingyun Sun, Papa Mao, Ying Zang
We can even outperform task-specific network models and achieve state-of-the-art performance in the task we tested: camouflaged object detection, shadow detection.
no code implementations • CVPR 2023 • Lanyun Zhu, Tianrun Chen, Jianxiong Yin, Simon See, Jun Liu
Continual Semantic Segmentation (CSS) extends static semantic segmentation by incrementally introducing new classes for training.
no code implementations • 4 Apr 2023 • Zhikang Liu, Lanyun Zhu
In this paper, we address this issue by designing a distillation method that exploits label structure when training segmentation network.
1 code implementation • 29 Mar 2022 • Xiao Fu, Shangzhan Zhang, Tianrun Chen, Yichong Lu, Lanyun Zhu, Xiaowei Zhou, Andreas Geiger, Yiyi Liao
In this work, we present a novel 3D-to-2D label transfer method, Panoptic NeRF, which aims for obtaining per-pixel 2D semantic and instance labels from easy-to-obtain coarse 3D bounding primitives.
1 code implementation • CVPR 2021 • Lanyun Zhu, Deyi Ji, Shiping Zhu, Weihao Gan, Wei Wu, Junjie Yan
In this paper, we fully take advantages of the low-level texture features and propose a novel Statistical Texture Learning Network (STLNet) for semantic segmentation.
1 code implementation • 27 Apr 2020 • Xuanyi Liu, Lanyun Zhu, Shiping Zhu, Li Luo
Recent works have achieved great success in improving the performance of multiple computer vision tasks by capturing features with a high channel number utilizing deep neural networks.
no code implementations • 17 Jul 2019 • Shiping Zhu, Lanyun Zhu
However, most of the existing attention modules used in salient object detection are input with the processed feature map itself, which easily leads to the problem of `blind overconfidence'.