Unknown

Dataset Information

0

Grouping and Segregation of Sensory Events by Actions in Temporal Audio-Visual Recalibration.


ABSTRACT: Perception in multi-sensory environments involves both grouping and segregation of events across sensory modalities. Temporal coincidence between events is considered a strong cue to resolve multisensory perception. However, differences in physical transmission and neural processing times amongst modalities complicate this picture. This is illustrated by cross-modal recalibration, whereby adaptation to audio-visual asynchrony produces shifts in perceived simultaneity. Here, we examined whether voluntary actions might serve as a temporal anchor to cross-modal recalibration in time. Participants were tested on an audio-visual simultaneity judgment task after an adaptation phase where they had to synchronize voluntary actions with audio-visual pairs presented at a fixed asynchrony (vision leading or vision lagging). Our analysis focused on the magnitude of cross-modal recalibration to the adapted audio-visual asynchrony as a function of the nature of the actions during adaptation, putatively fostering cross-modal grouping or, segregation. We found larger temporal adjustments when actions promoted grouping than segregation of sensory events. However, a control experiment suggested that additional factors, such as attention to planning/execution of actions, could have an impact on recalibration effects. Contrary to the view that cross-modal temporal organization is mainly driven by external factors related to the stimulus or environment, our findings add supporting evidence for the idea that perceptual adjustments strongly depend on the observer's inner states induced by motor and cognitive demands.

SUBMITTER: Ikumi N 

PROVIDER: S-EPMC5243829 | biostudies-literature | 2016

REPOSITORIES: biostudies-literature

altmetric image

Publications

Grouping and Segregation of Sensory Events by Actions in Temporal Audio-Visual Recalibration.

Ikumi Nara N   Soto-Faraco Salvador S  

Frontiers in integrative neuroscience 20170119


Perception in multi-sensory environments involves both grouping and segregation of events across sensory modalities. Temporal coincidence between events is considered a strong cue to resolve multisensory perception. However, differences in physical transmission and neural processing times amongst modalities complicate this picture. This is illustrated by cross-modal recalibration, whereby adaptation to audio-visual asynchrony produces shifts in perceived simultaneity. Here, we examined whether v  ...[more]

Similar Datasets

| S-EPMC4086723 | biostudies-literature
| S-EPMC10791168 | biostudies-literature
| S-EPMC5720774 | biostudies-literature
| S-EPMC6979315 | biostudies-literature
| S-EPMC5159801 | biostudies-literature
| S-EPMC3633943 | biostudies-literature
| S-EPMC8886615 | biostudies-literature
| S-EPMC8675747 | biostudies-literature
| S-EPMC4037709 | biostudies-other
| S-EPMC4831745 | biostudies-literature