1 code implementation • 31 Mar 2024 • Kun Ding, Haojian Zhang, Qiang Yu, Ying Wang, Shiming Xiang, Chunhong Pan
The idea is realized by exploiting out-of-distribution (OOD) detection to predict whether a sample belongs to a base distribution or a novel distribution and then using the score generated by a dedicated competition based scoring function to fuse the zero-shot and few-shot classifier.
no code implementations • 18 Mar 2024 • Kun Ding, Xiaohui Li, Qiang Yu, Ying Wang, Haojian Zhang, Shiming Xiang
Context Optimization (CoOp) has emerged as a simple yet effective technique for adapting CLIP-like vision-language models to downstream image recognition tasks.
1 code implementation • 29 Aug 2022 • Kun Ding, Ying Wang, Pengzhang Liu, Qiang Yu, Haojian Zhang, Shiming Xiang, Chunhong Pan
Inspired by the fact that modeling task relationship by multi-task learning can usually boost performance, we propose a novel method SoftCPT (Soft Context Sharing for Prompt Tuning) to tune pre-trained vision-language models on multiple target few-shot tasks jointly.
1 code implementation • IEEE Transactions on Cybernetics 2022 • Chenxiang Ma, Rui Yan, Zhaofei Yu, Qiang Yu
We then propose two variants that additionally incorporate temporal dependencies through a backward and forward process, respectively.
no code implementations • 2 Dec 2021 • Wenqiao Zhang, Haochen Shi, Siliang Tang, Jun Xiao, Qiang Yu, Yueting Zhuang
The contemporary visual captioning models frequently hallucinate objects that are not actually in a scene, due to the visual misclassification or over-reliance on priors that resulting in the semantic inconsistency between the visual information and the target lexical words.
no code implementations • 13 Apr 2021 • Zongshen Mu, Siliang Tang, Jie Tan, Qiang Yu, Yueting Zhuang
In this paper, we propose a novel graph learning framework for phrase grounding in the image.
Ranked #6 on Phrase Grounding on Flickr30k Entities Test
no code implementations • 11 May 2020 • Qiang Yu, Shiming Song, Chenxiang Ma, Linqiang Pan, Kay Chen Tan
Traditional neuron models use analog values for information representation and computation, while all-or-nothing spikes are employed in the spiking ones.
no code implementations • 5 May 2020 • Qiang Yu, Chenxiang Ma, Shiming Song, Gaoyan Zhang, Jianwu Dang, Kay Chen Tan
We examine the performance of our methods based on MNIST, Fashion-MNIST and CIFAR10 datasets.
no code implementations • 2 May 2020 • Qiang Yu, Shenglan Li, Huajin Tang, Longbiao Wang, Jianwu Dang, Kay Chen Tan
They are also believed to play an essential role in low-power consumption of the biological systems, whose efficiency attracts increasing attentions to the field of neuromorphic computing.
no code implementations • 4 Feb 2019 • Qiang Yu, Yanli Yao, Longbiao Wang, Huajin Tang, Jianwu Dang, Kay Chen Tan
Our framework is a unifying system with a consistent integration of three major functional parts which are sparse encoding, efficient learning and robust readout.