Perceptual learning of view-independence in visuo-haptic object representations

Simon Lacey, Marisa Pappas, Alexandra Kreps, Kevin Lee, K. Sathian

Research output: Contribution to journalArticlepeer-review

33 Scopus citations

Abstract

We previously showed that cross-modal recognition of unfamiliar objects is view-independent, in contrast to view-dependence within-modally, in both vision and haptics. Does the view-independent, bisensory representation underlying cross-modal recognition arise from integration of unisensory, view-dependent representations or intermediate, unisensory but view-independent representations? Two psychophysical experiments sought to distinguish between these alternative models. In both experiments, participants began from baseline, within-modal, view-dependence for object recognition in both vision and haptics. The first experiment induced within-modal view-independence by perceptual learning, which was completely and symmetrically transferred cross-modally: visual view-independence acquired through visual learning also resulted in haptic view-independence and vice versa. In the second experiment, both visual and haptic view-dependence were transformed to view-independence by either haptic-visual or visual-haptic cross-modal learning. We conclude that cross-modal view-independence fits with a model in which unisensory view-dependent representations are directly integrated into a bisensory, view-independent representation, rather than via intermediate, unisensory, view-independent representations.

Original languageEnglish (US)
Pages (from-to)329-337
Number of pages9
JournalExperimental Brain Research
Volume198
Issue number2-3
DOIs
StatePublished - Sep 2009

All Science Journal Classification (ASJC) codes

  • Neuroscience(all)

Fingerprint

Dive into the research topics of 'Perceptual learning of view-independence in visuo-haptic object representations'. Together they form a unique fingerprint.

Cite this