Language models outperform cloze predictability in a cognitive model of reading

Adrielli Tina Lopes Rego*, Joshua Snell, Martijn Meeter

*Corresponding author for this work

Research output: Contribution to JournalArticleAcademicpeer-review

Abstract

Although word predictability is commonly considered an important factor in reading, sophisticated accounts of predictability in theories of reading are lacking. Computational models of reading traditionally use cloze norming as a proxy of word predictability, but what cloze norms precisely capture remains unclear. This study investigates whether large language models (LLMs) can fill this gap. Contextual predictions are implemented via a novel parallel-graded mechanism, where all predicted words at a given position are pre-activated as a function of contextual certainty, which varies dynamically as text processing unfolds. Through reading simulations with OB1-reader, a cognitive model of word recognition and eye-movement control in reading, we compare the model’s fit to eye-movement data when using predictability values derived from a cloze task against those derived from LLMs (GPT-2 and LLaMA). Root Mean Square Error between simulated and human eye movements indicates that LLM predictability provides a better fit than cloze. This is the first study to use LLMs to augment a cognitive model of reading with higher-order language processing while proposing a mechanism on the interplay between word predictability and eye movements.

Original languageEnglish
Article numbere1012117
Pages (from-to)1-24
Number of pages24
JournalPLoS Computational Biology
Volume20
Issue number9
Early online date25 Sept 2024
DOIs
Publication statusPublished - Sept 2024

Bibliographical note

Publisher Copyright:
© 2024 Lopes Rego et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

Funding

This study was supported by the Nederlandse Organisatie voor Wetenschappelijk Onderzoek (NWO) Open Competition-SSH (Social Sciences and Humanities) (https://www.nwo.nl),406.21.GO.019 to MM. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

FundersFunder number
Nederlandse Organisatie voor Wetenschappelijk Onderzoek
Sociale en Geesteswetenschappen, NWO

    Fingerprint

    Dive into the research topics of 'Language models outperform cloze predictability in a cognitive model of reading'. Together they form a unique fingerprint.

    Cite this