Submission 495
The Representational Alignment Between Humans and Language Models Is Implicitly Driven by a Concreteness Effect
SymposiumTalk-02
Presented by: Cosimo Iaia
Words in human language can be conceptualized as more abstract (like justice) or more concrete (like table). Cognitive psychology has established that this property of word meaning influences how words are processed. Thus, understanding how concreteness is represented in our mind and brain is a central question in psychology, neuroscience, and computational linguistics. While the advent of powerful language models has allowed for quantitative inquiries into the nature of semantic representations, it remains largely underexplored how they represent concreteness. Here, we leveraged an odd-one-out task to estimate semantic distances implicitly used by humans, for a set of carefully selected abstract and concrete nouns. Using Representational Similarity Analysis, we find that this implicit representational space based on human ratings from 40 participants and the semantic representations of five frequently-used language models (fastText, word2vec, BERT base, BERT large, and GPT2) are significantly aligned, and that both representational spaces are aligned to an explicit representation of concreteness (based on concreteness ratings provided by the same participants). Most importantly, using model ablation experiments, we demonstrate that human-to-model alignment is substantially driven by concreteness, but not by other word characteristics like word length or frequency. Overall, human-to-model alignment is sensitive more to semantic than to non-semantic variables. Combined, these results highlight a critical role for concreteness as a mediator of the alignment between human word representations and language models. More generally, this work shows that language models can be useful tools for understanding the nature of semantic representations in the human mind and brain.