1 code implementation • 13 Mar 2024 • Cheng Chen, Junchen Zhu, Xu Luo, HengTao Shen, Lianli Gao, Jingkuan Song
To this end, we introduce MoELoRA to MLLMs which is effective to retain the previous instruction alignment.
1 code implementation • 16 Dec 2023 • Decheng Liu, Xu Luo, Chunlei Peng, Nannan Wang, Ruimin Hu, Xinbo Gao
In this paper, we propose a novel Symmetrical Bidirectional Knowledge Alignment for zero-shot sketch-based image retrieval (SBKA).
no code implementations • 15 Dec 2023 • Dingning Liu, Xiaomeng Dong, Renrui Zhang, Xu Luo, Peng Gao, Xiaoshui Huang, Yongshun Gong, Zhihui Wang
In this work, we present a new visual prompting method called 3DAxiesPrompts (3DAP) to unleash the capabilities of GPT-4V in performing 3D spatial tasks.
no code implementations • 5 Oct 2023 • Xu Luo, Difan Zou, Lianli Gao, Zenglin Xu, Jingkuan Song
Transferring a pretrained model to a downstream task can be as easy as conducting linear probing with target data, that is, training a linear classifier upon frozen features extracted from the pretrained model.
no code implementations • 4 Oct 2023 • Dun Zeng, Zenglin Xu, Yu Pan, Xu Luo, Qifan Wang, Xiaoying Tang
Central to this process is the technique of unbiased client sampling, which ensures a representative selection of clients.
no code implementations • 20 Jul 2023 • Zhipeng Zhang, Piao Tong, Yingwei Ma, Qiao Liu, Xujiang Liu, Xu Luo
Furthermore, we introduce a novel Decoupled Contrastive Learning method to enhance the effectiveness of the language representation.
2 code implementations • ICCV 2023 • Ji Zhang, Lianli Gao, Xu Luo, HengTao Shen, Jingkuan Song
Test-time task adaptation in few-shot learning aims to adapt a pre-trained task-agnostic model for capturing taskspecific knowledge of the test task, rely only on few-labeled support samples.
2 code implementations • 28 Jan 2023 • Xu Luo, Hao Wu, Ji Zhang, Lianli Gao, Jing Xu, Jingkuan Song
Few-shot classification consists of a training phase where a model is learned on a relatively large dataset and an adaptation phase where the learned model is adapted to previously-unseen tasks with limited labeled samples.
2 code implementations • 30 Oct 2022 • Jing Xu, Xu Luo, Xinglin Pan, Wenjie Pei, Yanan Li, Zenglin Xu
In this paper, we find that this problem usually occurs when the positions of support samples are in the vicinity of task centroid -- the mean of all class centroids in the task.
1 code implementation • 16 Jun 2022 • Xu Luo, Jing Xu, Zenglin Xu
When facing novel few-shot tasks in the test-time datasets, this transformation can greatly improve the generalization ability of learned image representations, while being agnostic to the choice of training algorithms and datasets.
no code implementations • 14 Dec 2021 • Jing Xu, Xinglin Pan, Xu Luo, Wenjie Pei, Zenglin Xu
To alleviate this problem, we present a simple yet effective feature rectification method by exploring the category correlation between novel and base classes as the prior knowledge.
no code implementations • 29 Sep 2021 • Shuhao Shi, Pengfei Xie, Xu Luo, Kai Qiao, Linyuan Wang, Jian Chen, Bin Yan
AMC-GNN generates two graph views by data augmentation and compares different layers' output embeddings of Graph Neural Network encoders to obtain feature representations, which could be used for downstream tasks.
1 code implementation • 20 Jul 2021 • Xu Luo, Yuxuan Chen, Liangjian Wen, Lili Pan, Zenglin Xu
The goal of few-shot classification is to classify new categories with few labeled examples within each class.
1 code implementation • NeurIPS 2021 • Xu Luo, Longhui Wei, Liangjian Wen, Jinrong Yang, Lingxi Xie, Zenglin Xu, Qi Tian
The category gap between training and evaluation has been characterised as one of the main obstacles to the success of Few-Shot Learning (FSL).
no code implementations • 1 Jan 2021 • Xu Luo, Yuxuan Chen, Liangjian Wen, Lili Pan, Zenglin Xu
Few-shot learning aims to recognize new classes with few annotated instances within each category.