Submission 702
The Role of Matching Versus Mismatching Granularity in Cross-Modal Event Memory
SymposiumTalk-03
Presented by: Tolgahan Aydın
Theories of multimodal learning, such as Dual Coding Theory, posit that aligning information across sensory channels is a key mechanism for building coherent event models. A direct prediction is that matching the granularity of information (e.g., fine- vs. coarse-grained details) between verbal and visual modalities should facilitate integration and enhance memory. We put this hypothesis to a test across two experiments. Participants viewed text and video presentations of everyday activities where granularity was systematically matched or mismatched. Recognition memory was assessed using a sensitive task designed to detect the benefits of alignment. Contrary to these theoretical foundations, we found no reliable benefit of cross-modal granularity matching on memory sensitivity, false alarms, or confidence. Bayesian analyses provided substantial evidence for this null effect. The critical factor was not alignment, but timing: memory was significantly superior when verbal descriptions preceded the video, scaffolding initial encoding. These findings pose a significant challenge to the assumption that structural alignment is a primary driver of multimodal integration in event memory. Instead, they demonstrate that the initial conditions of encoding—specifically, the order of presentation—can override the putative benefits of cross-modal congruence, forcing a reevaluation of how and when multimodal inputs effectively shape event models.