REVISTA
AI

   
Inicio  /  AI  /  Vol: 4 Par: 1 (2023)  /  Artículo
ARTÍCULO
TITULO

End-to-End Transformer-Based Models in Textual-Based NLP

Abir Rahali and Moulay A. Akhloufi    

Resumen

Transformer architectures are highly expressive because they use self-attention mechanisms to encode long-range dependencies in the input sequences. In this paper, we present a literature review on Transformer-based (TB) models, providing a detailed overview of each model in comparison to the Transformer?s standard architecture. This survey focuses on TB models used in the field of Natural Language Processing (NLP) for textual-based tasks. We begin with an overview of the fundamental concepts at the heart of the success of these models. Then, we classify them based on their architecture and training mode. We compare the advantages and disadvantages of popular techniques in terms of architectural design and experimental value. Finally, we discuss open research, directions, and potential future work to help solve current TB application challenges in NLP.

 Artículos similares