no code implementations • 20 Mar 2024 • Jinmin Li, Kuofeng Gao, Yang Bai, Jingyun Zhang, Shu-Tao Xia, Yisen Wang
Despite the remarkable performance of video-based large language models (LLMs), their adversarial threat remains unexplored.
1 code implementation • 23 Feb 2024 • Hang Guo, Jinmin Li, Tao Dai, Zhihao Ouyang, Xudong Ren, Shu-Tao Xia
In this way, our MambaIR takes advantage of the local pixel similarity and reduces the channel redundancy.
1 code implementation • 17 Dec 2023 • Yaohua Zha, Huizhen Ji, Jinmin Li, Rongsheng Li, Tao Dai, Bin Chen, Zhi Wang, Shu-Tao Xia
Specifically, to learn more compact features, a share-parameter Transformer encoder is introduced to extract point features from the global and local unmasked patches obtained by global random and local block mask strategies, followed by a specific decoder to reconstruct.
Ranked #3 on Few-Shot 3D Point Cloud Classification on ModelNet40 10-way (20-shot) (using extra training data)