Towards adaptive support for self-regulated learning of causal relations:Evaluating four Dutch word vector models

Advances in computational language models increasingly enable adaptive support for self-regulated learning (SRL) in digital learning environments (DLEs; eg, via automated feedback). However, the accuracy of those models is a common concern for educational stakeholders (eg, policymakers, researchers, teachers and learners themselves). We compared the accuracy of four Dutch language models (ie, spaCy medium, spaCy large, FastText and ConceptNet NumberBatch) in the context of secondary school students' learning of causal relations from expository texts, scaffolded by causal diagram completion. Si... Mehr ...

Verfasser: Pijeira-Diaz, Hector J.
Braumann, Sophia
van de Pol, Janneke
van Gog, Tamara
de Bruin, Anique B. H.
Dokumenttyp: Artikel
Erscheinungsdatum: 2024
Reihe/Periodikum: Pijeira-Diaz , H J , Braumann , S , van de Pol , J , van Gog , T & de Bruin , A B H 2024 , ' Towards adaptive support for self-regulated learning of causal relations : Evaluating four Dutch word vector models ' , British Journal of Educational Technology . https://doi.org/10.1111/bjet.13431
Schlagwörter: automatic scoring evaluation / causal relation learning / computational linguistic models / diagram completion task / semantic similarity
Sprache: Englisch
Permalink: https://search.fid-benelux.de/Record/base-27440537
Datenquelle: BASE; Originalkatalog
Powered By: BASE
Link(s) : https://cris.maastrichtuniversity.nl/en/publications/e9193e3b-4b9b-47ef-ba52-33a05c8d68b0

Advances in computational language models increasingly enable adaptive support for self-regulated learning (SRL) in digital learning environments (DLEs; eg, via automated feedback). However, the accuracy of those models is a common concern for educational stakeholders (eg, policymakers, researchers, teachers and learners themselves). We compared the accuracy of four Dutch language models (ie, spaCy medium, spaCy large, FastText and ConceptNet NumberBatch) in the context of secondary school students' learning of causal relations from expository texts, scaffolded by causal diagram completion. Since machine learning relies on human-labelled data for the best results, we used a dataset with 10,193 students' causal diagram answers, compiled over a decade of research using a diagram completion intervention to enhance students' monitoring of their text comprehension. The language models were used in combination with four popular machine learning classifiers (ie, logistic regression, random forests, support vector machine and neural networks) to evaluate their performance on automatically scoring students' causal diagrams in terms of the correctness of events and their sequence (ie, the causal structure). Five performance metrics were studied, namely accuracy, precision, recall, F1 and the area under the curve of the receiver operating characteristic (ROC-AUC). The spaCy medium model combined with the neural network classifier achieved the best performance for the correctness of causal events in four of the five metrics, while the ConceptNet NumberBatch model worked best for the correctness of the causal sequence. These evaluation results provide a criterion for model adoption to adaptively support SRL of causal relations in DLEs.Practitioner notesWhat is already known about this topicAccurate monitoring is a prerequisite for effective self-regulation.Students struggle to accurately monitor their comprehension of causal relations in texts.Completing causal diagrams improves students' monitoring accuracy, but there is ...