Submission 505
Beyond Lemmas: Modeling the Picture in Picture-Word Interference
SymposiumTalk-02
Presented by: Louis Schiekiera
Accounts of picture–word interference (PWI) typically assume that the picture has already been conceptually encoded and thus model interference exclusively at the conceptual–lexical level. In contrast, we test whether visual properties of the target picture itself contribute to interference effects in naming. We assembled a large multimodal PWI dataset— 189,767 trials from 23 experiments across 13 studies, involving more than 1,078 participants and 1,311 target images (with additional data still being collected). For each target picture and distractor word, we compute vision–language embeddings using OpenCLIP, deriving cosine-similarity measures for (a) image × distractor-word, (b) image × target-word, and (c) target-word × distractor-word pairs. These multimodal similarity estimates are integrated with trial-level behavioral data in mixed-effects models predicting log naming latencies. The project aims to assess whether image–word similarity provides explanatory power beyond traditional semantic relatedness measures and design factors. More broadly, the study offers a framework for incorporating computational visual representations into psycholinguistic models of word production.