2 code implementations • 23 Apr 2021 • Jochen Zöllner, Konrad Sperfeld, Christoph Wick, Roger Labahn
Currently, the most widespread neural network architecture for training language models is the so called BERT which led to improvements in various Natural Language Processing (NLP) tasks.