What Would Elsa Do? Freezing Layers During Transformer Fine-Tuning

8 Nov 2019 Jaejun Lee Raphael Tang Jimmy Lin

Pretrained transformer-based language models have achieved state of the art across countless tasks in natural language processing. These models are highly expressive, comprising at least a hundred million parameters and a dozen layers... (read more)

PDF Abstract

Results from the Paper

  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods used in the Paper