Redirigiendo al acceso original de articulo en 20 segundos...
Inicio  /  Future Internet  /  Vol: 15 Par: 1 (2023)  /  Artículo
ARTÍCULO
TITULO

BART-IT: An Efficient Sequence-to-Sequence Model for Italian Text Summarization

Moreno La Quatra and Luca Cagliero    

Resumen

The emergence of attention-based architectures has led to significant improvements in the performance of neural sequence-to-sequence models for text summarization. Although these models have proved to be effective in summarizing English-written documents, their portability to other languages is limited thus leaving plenty of room for improvement. In this paper, we present BART-IT, a sequence-to-sequence model, based on the BART architecture that is specifically tailored to the Italian language. The model is pre-trained on a large corpus of Italian-written pieces of text to learn language-specific features and then fine-tuned on several benchmark datasets established for abstractive summarization. The experimental results show that BART-IT outperforms other state-of-the-art models in terms of ROUGE scores in spite of a significantly smaller number of parameters. The use of BART-IT can foster the development of interesting NLP applications for the Italian language. Beyond releasing the model to the research community to foster further research and applications, we also discuss the ethical implications behind the use of abstractive summarization models.

 Artículos similares