Scene context automatically drives predictions of object transformations

Giacomo Aldegheri*, Surya Gayet, Marius V. Peelen

*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

As our viewpoint changes, the whole scene around us rotates coherently. This allows us to predict how one part of a scene (e.g., an object) will change by observing other parts (e.g., the scene background). While human object perception is known to be strongly context-dependent, previous research has largely focused on how scene context can disambiguate fixed object properties, such as identity (e.g., a car is easier to recognize on a road than on a beach). It remains an open question whether object representations are updated dynamically based on the surrounding scene context, for example across changes in viewpoint. Here, we tested whether human observers dynamically and automatically predict the appearance of objects based on the orientation of the background scene. In three behavioral experiments (N = 152), we temporarily occluded objects within scenes that rotated. Upon the objects' reappearance, participants had to perform a perceptual discrimination task, which did not require taking the scene rotation into account. Performance on this orthogonal task strongly depended on whether objects reappeared rotated coherently with the surrounding scene or not. This effect persisted even when a majority of trials violated this real-world contingency between scene and object, showcasing the automaticity of these scene-based predictions. These findings indicate that contextual information plays an important role in predicting object transformations in structured real-world environments.

Original languageEnglish
Article number105521
Number of pages9
JournalCognition
Volume238
DOIs
Publication statusPublished - Sept 2023

Keywords

  • Mental rotation
  • Object perception
  • Scene perception
  • Visual expectations

Fingerprint

Dive into the research topics of 'Scene context automatically drives predictions of object transformations'. Together they form a unique fingerprint.

Cite this