Inicio  /  Future Internet  /  Vol: 16 Par: 1 (2024)  /  Artículo
ARTÍCULO
TITULO

Evaluating Embeddings from Pre-Trained Language Models and Knowledge Graphs for Educational Content Recommendation

Xiu Li    
Aron Henriksson    
Martin Duneld    
Jalal Nouri and Yongchao Wu    

Resumen

Educational content recommendation is a cornerstone of AI-enhanced learning. In particular, to facilitate navigating the diverse learning resources available on learning platforms, methods are needed for automatically linking learning materials, e.g., in order to recommend textbook content based on exercises. Such methods are typically based on semantic textual similarity (STS) and the use of embeddings for text representation. However, it remains unclear what types of embeddings should be used for this task. In this study, we carry out an extensive empirical evaluation of embeddings derived from three different types of models: (i) static embeddings trained using a concept-based knowledge graph, (ii) contextual embeddings from a pre-trained language model, and (iii) contextual embeddings from a large language model (LLM). In addition to evaluating the models individually, various ensembles are explored based on different strategies for combining two models in an early vs. late fusion fashion. The evaluation is carried out using digital textbooks in Swedish for three different subjects and two types of exercises. The results show that using contextual embeddings from an LLM leads to superior performance compared to the other models, and that there is no significant improvement when combining these with static embeddings trained using a knowledge graph. When using embeddings derived from a smaller language model, however, it helps to combine them with knowledge graph embeddings. The performance of the best-performing model is high for both types of exercises, resulting in a mean Recall@3 of 0.96 and 0.95 and a mean MRR of 0.87 and 0.86 for quizzes and study questions, respectively, demonstrating the feasibility of using STS based on text embeddings for educational content recommendation. The ability to link digital learning materials in an unsupervised manner?relying only on readily available pre-trained models?facilitates the development of AI-enhanced learning.

 Artículos similares

       
 
André Melo, Btissam Er-Rahmadi and Jeff Z. Pan    
Aligning points of interest (POIs) from heterogeneous geographical data sources is an important task that helps extend map data with information from different datasets. This task poses several challenges, including differences in type hierarchies, label... ver más

 
Andrey Bogdanchikov, Dauren Ayazbayev and Iraklis Varlamis    
The rapid development of natural language processing and deep learning techniques has boosted the performance of related algorithms in several linguistic and text mining tasks. Consequently, applications such as opinion mining, fake news detection or doc... ver más

 
Guizhe Song and Degen Huang    
The massive amount of data generated by social media present a unique opportunity for disaster analysis. As a leading social platform, Twitter generates over 500 million Tweets each day. Due to its real-time characteristic, more agencies employ Twitter t... ver más
Revista: Future Internet

 
Dandan Li and Douglas Summers-Stay    
Word embeddings have been very successful in many natural language processing tasks, but they characterize the meaning of a word/concept by uninterpretable ?context signatures?. Such a representation can render results obtained using embeddings difficult... ver más