1 code implementation • 28 Mar 2024 • Wei Dong, Xing Zhang, Bihui Chen, Dawei Yan, Zhijun Lin, Qingsen Yan, Peng Wang, Yang Yang
Parameter-efficient fine-tuning for pre-trained Vision Transformers aims to adeptly tailor a model to downstream tasks by learning a minimal set of new adaptation parameters while preserving the frozen majority of pre-trained parameters.
1 code implementation • NeurIPS 2023 • Wei Dong, Dawei Yan, Zhijun Lin, Peng Wang
Consequently, effectively adapting large pre-trained models to downstream tasks in an efficient manner has become a prominent research area.
1 code implementation • 9 Feb 2023 • Wei Dong, Dawei Yan, Peng Wang
Considering the excessive memory overheads of contrastive learning, we further propose a negative-free solution, where the main contribution is a Graph Signal Decorrelation (GSD) constraint to avoid representation collapse and over-smoothing.
no code implementations • 7 Dec 2021 • Binglu Wang, Chengzhe Lu, Dawei Yan, Yongqiang Zhao
Secondly, as neighboring pixels exhibit different affinities to the center pixel, we estimate a local attention mask to modulate the influence of neighboring pixels.
no code implementations • 9 Dec 2019 • Dawei Yan, Shri Yadav, Andrea Paterlini, William Nicolas, Ilya Belevich, Magali Grison, Anne Vaten, Leila Karami, Sedeer El-Showk, Jung-Youn Lee, Gosia Murawska, Jenny Mortimer, Michael Knoblauch, Eija Jokitalo, Jonathan Markham, Emmanuelle Bayer, Ykä Helariutta
During phloem unloading, multiple cell-to-cell transport events move organic substances to the root meristem.