11:00 - 12:30
Parallel sessions 5
11:00 - 12:30
Room: HSZ - N2
Chair/s:
Cosimo Iaia, Jack E Taylor
The development of powerful computational language models in recent years has seen an increasing application of such models in psychology and psycholinguistics. Both Distributional Semantic Models (e.g., word2vec, GloVe, etc) and Large Language Models (e.g., GPT, BERT) have been applied in two main ways: (1) as models of human language processing, and (2) as tools for generating measures that are relevant to psycholinguistic theories and hypotheses. However, the distinction between these two applications of language models is not always clear, and both applications are limited by fundamental differences in language processing between models engineered to achieve human-like performance, and the processes actually used in human language. Nevertheless, language models have demonstrated strong potential for providing insight into language processes. This symposium brings together five talks to address recent developments in the use of language models as tools for psycholinguistics, and the degree to which such models provide comparisons and outputs that are meaningful for progress in the field. The first talk will set a theoretical foundation for the symposium, evaluating caveats of comparing Large Language Models to humans, and outlining how meaningful comparisons require rigorous experimental methods. The second talk explores whether humans and language models (both Large Language Models and Distributional Semantic Models) represent abstract meaning in a similar way, while also highlighting differences emerging between the two systems. The third talk shows how Large Language Models can be used to generate new iconicity ratings for Turkish, providing a new avenue for investigating semantic dimensions in otherwise understudied languages. The fourth talk evaluates how well estimates of word frequency and familiarity derived from Large Language Models can explain children’s reading times. Finally, in the last talk, Distributional Semantic Models are applied to provide insight into the learning of morphology, showing that natural text provides sufficient information to learn the meanings of prefixes and suffixes. Together, these talks highlight the ongoing potential of language models as tools for psycholinguistics. However, these talks also provide opportunity for important discussion on the caveats of this approach, and on the scientific applications language models can support.
Submission 107
Modelling Affix Learning from Reading: Insights from Compositional Distributional Semantics
SymposiumTalk-05
Presented by: Maria Korochkina
Maria Korochkina 1, Marco Marelli 2, Kathleen Rastle 1
1 Royal Holloway, University of London, United Kingdom
2 University of Milano-Bicocca, Italy
Most words in English and other languages are built by combining smaller units of meaning called morphemes (e.g., teach + -erteacher). Understanding a language’s morphology is essential for developing reading expertise because it enables readers to compute the meanings of familiar and unfamiliar words (e.g., tweeter). Our work investigated what English-speaking children can learn about English morphology through text experience. We trained a compositional distributional semantic model on words from 1,200 books popular with British children aged 7–16 to investigate what the model can learn about affix meanings through reading. We then assessed whether the model’s knowledge of affix meanings aligned with that of 120 adults in a lexical decision task. The model’s knowledge of individual affixes accounted for patterns in readers’ lexical processing: affixes estimated by the model to have richer, more coherent meanings were better known, and this was associated with both the number of distinct words containing each affix and the level of noise in their usage. Our work shows that, despite high levels of noise, natural text contains enough structure to support the extraction of core affix semantics, and that readers’ knowledge of affixes reflects the patterns found in text. This study adds a new dimension to a more principled and psychologically grounded understanding of morpheme learning.