Multi-Task Learning with Auxiliary Speaker Identification for Conversational Emotion Recognition

3 Mar 2020  ·  Jingye Li, Meishan Zhang, Donghong Ji, Yijiang Liu ·

Conversational emotion recognition (CER) has attracted increasing interests in the natural language processing (NLP) community. Different from the vanilla emotion recognition, effective speaker-sensitive utterance representation is one major challenge for CER. In this paper, we exploit speaker identification (SI) as an auxiliary task to enhance the utterance representation in conversations. By this method, we can learn better speaker-aware contextual representations from the additional SI corpus. Experiments on two benchmark datasets demonstrate that the proposed architecture is highly effective for CER, obtaining new state-of-the-art results on two datasets.

PDF Abstract

Datasets


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Emotion Recognition in Conversation EmoryNLP BERT+MTL Weighted-F1 35.92 # 22
Emotion Recognition in Conversation EmoryNLP GloVE+MTL Weighted-F1 34.54 # 24
Emotion Recognition in Conversation MELD BERT+MTL Weighted-F1 61.90 # 42
Emotion Recognition in Conversation MELD GloVE+MTL Weighted-F1 60.69 # 47

Methods


No methods listed for this paper. Add relevant methods here