no code implementations • 15 Jan 2024 • Yi Zhang, Ce Zhang, Ke Yu, Yushun Tang, Zhihai He
However, for generalization tasks, the current fine-tuning methods for CLIP, such as CoOp and CoCoOp, demonstrate relatively low performance on some fine-grained datasets.
no code implementations • 3 Sep 2023 • Yi Zhang, Ce Zhang, Zihan Liao, Yushun Tang, Zhihai He
Large-scale pre-trained Vision-Language Models (VLMs), such as CLIP and ALIGN, have introduced a new paradigm for learning transferable visual representations.
no code implementations • 28 Jul 2023 • Yi Zhang, Ce Zhang, Yushun Tang, Zhihai He
Based on these visual concepts, we construct a discriminative representation of images and learn a concept inference network to perform downstream image classification tasks, such as few-shot learning and domain generalization.
no code implementations • 29 Jun 2023 • Yushun Tang, Qinghai Guo, Zhihai He
Our main idea is that, when we adapt the network model to predict the sample labels from encoded features, we use these prediction results to construct new training samples with derived labels to learn a new examiner network that performs a different but compatible task in the target domain.
no code implementations • CVPR 2023 • Zhehan Kan, Shuoshuo Chen, Ce Zhang, Yushun Tang, Zhihai He
This strong correlation suggests that we can use this error as feedback to guide the correction process.
no code implementations • CVPR 2023 • Yushun Tang, Ce Zhang, Heng Xu, Shuoshuo Chen, Jie Cheng, Luziwei Leng, Qinghai Guo, Zhihai He
We observe that the performance of this feed-forward Hebbian learning for fully test-time adaptation can be significantly improved by incorporating a feedback neuro-modulation layer.