2 code implementations • 24 Aug 2023 • Kevin Krahn, Derrick Tate, Andrew C. Lamicela
In this work, we use a multilingual knowledge distillation approach to train BERT models to produce sentence embeddings for Ancient Greek text.
Authorship Attribution Knowledge Distillation +11