Error-correction mechanisms in language learning: modeling individuals

Abstract

Since its first adoption as a computational model for language learning, evidence has accumulated that Rescorla–Wagner error-correction learning (Rescorla & Wagner, 1972) captures several aspects of language processing. Whereas previous studies have provided general support for the Rescorla–Wagner rule by using it to explain the behavior of participants across a range of tasks, we focus on testing predictions generated by the model in a controlled natural language learning task and model the data at the level of the individual learner. By adjusting the parameters of the model to fit the trial-by-trial behavioral choices of participants, rather than fitting a one-for-all model using a single set of default parameters, we show that the model accurately captures participants’ choices, time latencies, and levels of response agreement. We also show that gender and working memory capacity affect the extent to which the Rescorla–Wagner model captures language learning.

Publication
Language Learning
Avatar
Adnane Ez-zizi
Senior Lecturer in Artificial Intelligence

My research interests include reinforcement learning, educational data mining and AI, Natural language processing and computational modelling of human behaviour.