BART-IT: An Efficient Sequence-to-Sequence Model for Italian Text Summarization

The emergence of attention-based architectures has led to significant improvements in the performance of neural sequence-to-sequence models for text summarization. Although these models have proved to be effective in summarizing English-written documents, their portability to other languages is limited thus leaving plenty of room for improvement. In this paper, we present BART-IT, a sequence-to-sequence model, based on the BART architecture that is specifically tailored to the Italian language. The model is pre-trained on a large corpus of Italian-written pieces of text to learn language-specific features and then fine-tuned on several benchmark datasets established for abstractive summarization. The experimental results show that BART-IT outperforms other state-of-the-art models in terms of ROUGE scores in spite of a significantly smaller number of parameters. The use of BART-IT can foster the development of interesting NLP applications for the Italian language. Beyond releasing the model to the research community to foster further research and applications, we also discuss the ethical implications behind the use of abstractive summarization models.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Abstractive Text Summarization Abstractive Text Summarization from Fanpage mBART ROUGE-1 36.52 # 1
# Parameters 610 # 1
ROUGE-2 17.52 # 1
ROUGE-L 26.14 # 1
BERTScore 73.4 # 1
Abstractive Text Summarization Abstractive Text Summarization from Fanpage mT5 ROUGE-1 34.13 # 4
# Parameters 390 # 2
ROUGE-2 15.76 # 3
ROUGE-L 24.84 # 4
BERTScore 72.77 # 3
Abstractive Text Summarization Abstractive Text Summarization from Fanpage BART-IT ROUGE-1 35.42 # 3
# Parameters 140 # 4
ROUGE-2 15.88 # 2
ROUGE-L 25.12 # 2
BERTScore 73.24 # 2
Abstractive Text Summarization Abstractive Text Summarization from Fanpage IT5-base ROUGE-1 33.99 # 5
# Parameters 220 # 3
ROUGE-2 15.59 # 4
ROUGE-L 24.91 # 3
BERTScore 70.3 # 4
Abstractive Text Summarization Abstractive Text Summarization from Il Post IT5-base ROUGE-1 32.88 # 6
ROUGE-2 15.53 # 4
ROUGE-L 26.7 # 4
BERTScore 71.06 # 4
Abstractive Text Summarization Abstractive Text Summarization from Il Post mT5 ROUGE-1 35.04 # 4
ROUGE-2 17.41 # 3
ROUGE-L 28.68 # 3
BERTScore 74.69 # 3
Abstractive Text Summarization Abstractive Text Summarization from Il Post BART-IT ROUGE-1 37.31 # 3
ROUGE-2 19.44 # 2
ROUGE-L 30.41 # 2
BERTScore 75.36 # 2
Abstractive Text Summarization Abstractive Text Summarization from Il Post mBART ROUGE-1 38.91 # 1
ROUGE-2 21.41 # 1
ROUGE-L 32.08 # 1
BERTScore 75.86 # 1
Abstractive Text Summarization WITS BART-IT ROUGE-1 42.32 # 1
ROUGE-2 28.83 # 1
ROUGE-L 38.84 # 1
BERTScore 79.28 # 2
Abstractive Text Summarization WITS mT5 ROUGE-1 40.6 # 2
ROUGE-2 26.9 # 2
ROUGE-L 37.43 # 2
BERTScore 80.73 # 1
Abstractive Text Summarization WITS mBART ROUGE-1 39.32 # 3
ROUGE-2 26.18 # 3
ROUGE-L 35.9 # 3
BERTScore 78.65 # 3
Abstractive Text Summarization WITS IT5-base ROUGE-1 37.98 # 4
ROUGE-2 24.32 # 4
ROUGE-L 34.94 # 4
BERTScore 77.14 # 4

Methods