Search Results for author: Haibo Su

Found 2 papers, 1 papers with code

To Tune or Not To Tune? How About the Best of Both Worlds?

3 code implementations9 Jul 2019 Ran Wang, Haibo Su, Chunye Wang, Kailin Ji, Jupeng Ding

In this regard, Peters et al. perform several experiments which demonstrate that it is better to adapt BERT with a light-weight task-specific head, rather than building a complex one on top of the pre-trained language model, and freeze the parameters in the said language model.

Language Modelling Semantic Similarity +2

Cannot find the paper you are looking for? You can Submit a new open access paper.