A Hierarchical Multi-task Approach for Learning Embeddings from Semantic Tasks

14 Nov 2018  ·  Victor Sanh, Thomas Wolf, Sebastian Ruder ·

Much effort has been devoted to evaluate whether multi-task learning can be leveraged to learn rich representations that can be used in various Natural Language Processing (NLP) down-stream applications. However, there is still a lack of understanding of the settings in which multi-task learning has a significant effect. In this work, we introduce a hierarchical model trained in a multi-task learning setup on a set of carefully selected semantic tasks. The model is trained in a hierarchical fashion to introduce an inductive bias by supervising a set of low level tasks at the bottom layers of the model and more complex tasks at the top layers of the model. This model achieves state-of-the-art results on a number of tasks, namely Named Entity Recognition, Entity Mention Detection and Relation Extraction without hand-engineered features or external NLP tools like syntactic parsers. The hierarchical training supervision induces a set of shared semantic representations at lower layers of the model. We show that as we move from the bottom to the top layers of the model, the hidden states of the layers tend to represent more complex semantic information.

PDF Abstract

Results from the Paper


Ranked #10 on Relation Extraction on ACE 2005 (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Relation Extraction ACE 2005 Hierarchical Multi-task RE Micro F1 62.7 # 10
NER Micro F1 87.5 # 11
Sentence Encoder ELMo # 1
Cross Sentence No # 1

Methods


No methods listed for this paper. Add relevant methods here