no code implementations • 26 Feb 2024 • Anchun Gui, Jian Li, Yong Dai, Nan Du, Han Xiao
Meanwhile, we propose a novel tool sampling strategy to enhance the generalizability of LLMs over unseen tools.
no code implementations • 17 May 2023 • Anchun Gui, Jinqiang Ye, Han Xiao
However, with the growth of model scale and the rising number of downstream tasks, this paradigm inevitably meets the challenges in terms of computation consumption and memory footprint issues.
no code implementations • 8 May 2023 • Anchun Gui, Han Xiao
To fully leverage the advantages of large-scale pre-trained language models (PLMs) on downstream tasks, it has become a ubiquitous adaptation paradigm to fine-tune the entire parameters of PLMs.