Supervised Adversarial Contrastive Learning for Emotion Recognition in Conversations

2 Jun 2023  ·  Dou Hu, Yinan Bao, Lingwei Wei, Wei Zhou, Songlin Hu ·

Extracting generalized and robust representations is a major challenge in emotion recognition in conversations (ERC). To address this, we propose a supervised adversarial contrastive learning (SACL) framework for learning class-spread structured representations in a supervised manner. SACL applies contrast-aware adversarial training to generate worst-case samples and uses joint class-spread contrastive learning to extract structured representations. It can effectively utilize label-level feature consistency and retain fine-grained intra-class features. To avoid the negative impact of adversarial perturbations on context-dependent data, we design a contextual adversarial training (CAT) strategy to learn more diverse features from context and enhance the model's context robustness. Under the framework with CAT, we develop a sequence-based SACL-LSTM to learn label-consistent and context-robust features for ERC. Experiments on three datasets show that SACL-LSTM achieves state-of-the-art performance on ERC. Extended experiments prove the effectiveness of SACL and CAT.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Emotion Recognition in Conversation EmoryNLP SACL-LSTM (one seed) Weighted-F1 40.47 # 4
Micro-F1 43.19 # 2
Emotion Recognition in Conversation EmoryNLP SACL-LSTM Weighted-F1 39.65 # 8
Micro-F1 42.21 # 5
Emotion Recognition in Conversation IEMOCAP SACL-LSTM (one seed) Weighted-F1 69.70 # 15
Accuracy 69.62 # 10
Emotion Recognition in Conversation IEMOCAP SACL-LSTM Weighted-F1 69.22 # 18
Accuracy 69.08 # 13
Emotion Recognition in Conversation MELD SACL-LSTM (one seed) Weighted-F1 66.86 # 8
Accuracy 67.89 # 2
Emotion Recognition in Conversation MELD SACL-LSTM Weighted-F1 66.45 # 16
Accuracy 67.51 # 6

Methods