Abstract
Lexical processing times can yield valuable insights about structure in language and the cognitive processes that enable the use of language. Being able to estimate lexical processing times enables us to estimate readability and reading times of any text. It has been claimed that lexical processing times of words are influenced by word occurrence frequencies as well as the context it appears in (McDonald and Shillcock 2001; Baayen 2010). The context might be important because of predictive processes that enable quicker lexical processing (Christiansen and Chater 2016). In the present paper, the effects of morphosyntactic predictions on lexical processing times are investigated using two computational models. These computational models are trained to predict upcoming part-of-speech tags based on preceding part-of-speech tags and their predictions are compared with human predictions and human reading times from the PROVO corpus (Luke and Christianson 2018). A recurrent neural network is able to explain variance in human prediction errors whereas the Rescorla-Wagner model performs less well. The Rescorla-Wagner prediction associations do however explain more variance in human reading times. Moreover, the Rescorla-Wagner model associations explain more variance in gaze durations than human prediction errors. The human prediction errors and the Recorla-Wagner model associations combined explain most variance (Adj. R\(^2 = 0.719\)) in reading times, which indicates that the part-of-speech tag-based Rescorla-Wagner model associations contain complementary information to explicit human predictions about lexical processing times.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Baayen, R.H.: Demythologizing the word frequency effect: a discriminative learning perspective. Mental Lexicon 5(3), 436–461 (2010)
Baayen, R.H., Milin, P., Durdjevic, D.F., Hendrix, P., Marelli, M.: An amorphous model for morphological processing in visual comprehension based on Naive discriminative learning. Psychol. Rev. 118(3), 438 (2011)
Balota, D.A., Yap, M.J., Hutchison, K.A., Cortese, M.J., Kessler, B., Loftis, B., et al.: The English lexicon project. Behav. Res. Methods 39(3), 445–459 (2007)
Christiansen, M.H., Chater, N.: The now-or-never bottleneck: a fundamental constraint on language. Behav. Brain Sci. 39 (2016)
Elman, J.L.: Finding structure in time. Cogn. Sci. 14(2), 179–211 (1990)
Garside, R.: A hybrid grammatical tagger: Claws 4, Corpus annotation. Linguistic information from computer text corpora (1997)
Hinojosa, J.A., Moreno, E.M., Casado, P., Muñoz, F., Pozo, M.A.: Syntactic expectancy: an event-related potentials study. Neurosci. Lett. 378(1), 34–39 (2005)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization (2014). arXiv preprint arXiv:1412.6980
Loper, E., Bird, S.: NLTK: the natural language toolkit. arXiv preprint cs/0205028 (2002)
Luke, S.G., Christianson, K.: Limits on lexical prediction during reading. Cogn. Psychol. 88, 22–60 (2016)
Luke, S.G., Christianson, K.: The Provo corpus: a large eye-tracking corpus with predictability norms. Behav. Res. Methods 50, 826–833 (2018)
Marcus, M., Santorini, B., Marcinkiewicz, M.A.: Building a large annotated corpus of English: The Penn Treebank (1993)
McClelland, J.L., Rumelhart, D.E.: An interactive activation model of context effects in letter perception: I. An account of basic findings. Psychol. Rev. 88(5), 375 (1981)
McDonald, S.A., Shillcock, R.C.: Rethinking the word frequency effect: the neglected role of distributional information in lexical processing. Lang. Speech 44(3), 295–322 (2001)
Mirman, D., Graf Estes, K., Magnuson, J.S.: Computational modeling of statistical learning: effects of transitional probability versus frequency and links to word learning. Infancy 15(5), 471–486 (2010)
Misyak, J.B., Christiansen, M.H., Bruce Tomblin, J.: Sequential expectations: the role of prediction-based learning in language. Topics Cogn. Sci. 2(1), 138–153 (2010)
Nivre, J., Abrams, M., Agić, Ž., et al.: Universal dependencies 2.3. (LINDAT/CLARIN digital library at the Institute of Formal and Applied Linguistics (ÚFAL), Faculty of Mathematics and Physics, Charles University) (2018)
Rescorla, R.A., Wagner, A.R., et al.: A theory of Pavlovian conditioning: variations in the effectiveness of reinforcement and nonreinforcement. Class. Conditioning II Curr. Res. Theory 2, 64–99 (1972)
Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929–1958 (2014)
Taylor, W.L.: “Cloze procedure”: a new tool for measuring readability. J. Bull. 30(4), 415–433 (1953)
Van Berkum, J.J., Brown, C.M., Zwitserlood, P., Kooijman, V., Hagoort, P.: Anticipating upcoming words in discourse: evidence from ERPs and reading times. J. Exp. Psychol. Learn. Mem. Cognit. 31(3), 443 (2005)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
de Vries, W. (2020). Cognitively Plausible Computational Models of Lexical Processing Can Explain Variance in Human Word Predictions and Reading Times. In: Bogaerts, B., et al. Artificial Intelligence and Machine Learning. BNAIC BENELEARN 2019 2019. Communications in Computer and Information Science, vol 1196. Springer, Cham. https://doi.org/10.1007/978-3-030-65154-1_4
Download citation
DOI: https://doi.org/10.1007/978-3-030-65154-1_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-65153-4
Online ISBN: 978-3-030-65154-1
eBook Packages: Computer ScienceComputer Science (R0)