HateXplain: A Benchmark Dataset for Explainable Hate Speech Detection

Hate speech is a challenging issue plaguing the online social media. While better models for hate speech detection are continuously being developed, there is little research on the bias and interpretability aspects of hate speech. In this paper, we introduce HateXplain, the first benchmark hate speech dataset covering multiple aspects of the issue. Each post in our dataset is annotated from three different perspectives: the basic, commonly used 3-class classification (i.e., hate, offensive or normal), the target community (i.e., the community that has been the victim of hate speech/offensive speech in the post), and the rationales, i.e., the portions of the post on which their labelling decision (as hate, offensive or normal) is based. We utilize existing state-of-the-art models and observe that even models that perform very well in classification do not score high on explainability metrics like model plausibility and faithfulness. We also observe that models, which utilize the human rationales for training, perform better in reducing unintended bias towards target communities. We have made our code and dataset public at https://github.com/punyajoy/HateXplain

PDF Abstract

Datasets


Introduced in the Paper:

HateXplain

Used in the Paper:

Hate Speech
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Hate Speech Detection HateXplain BERT-HateXplain [Attn] AUROC 0.851 # 3
Accuracy 0.698 # 3
Macro F1 0.687 # 3
Hate Speech Detection HateXplain BERT-HateXplain [LIME] AUROC 0.851 # 3
Macro F1 0.687 # 3
Hate Speech Detection HateXplain BERT [Attn] AUROC 0.843 # 5
Accuracy 0.69 # 4
Macro F1 0.674 # 5
Hate Speech Detection HateXplain BiRNN-HateXplain [Attn] AUROC 0.805 # 6
Macro F1 0.629 # 6
Hate Speech Detection HateXplain BiRNN-Attn [Attn] AUROC 0.795 # 7
Accuracy 0.621 # 6
Hate Speech Detection HateXplain BiRNN [LIME] AUROC 0.767 # 9
Accuracy 0.595 # 7
Macro F1 0.575 # 8
Hate Speech Detection HateXplain CNN-GRU [LIME] AUROC 0.793 # 8
Accuracy 0.629 # 5
Macro F1 0.614 # 7

Methods