Article ID Journal Published Year Pages File Type
10458109 Cognition 2005 10 Pages PDF
Abstract
In face-to-face conversation speech is perceived by ear and eye. We studied the prerequisites of audio-visual speech perception by using perceptually ambiguous sine wave replicas of natural speech as auditory stimuli. When the subjects were not aware that the auditory stimuli were speech, they showed only negligible integration of auditory and visual stimuli. When the same subjects learned to perceive the same auditory stimuli as speech, they integrated the auditory and visual stimuli in a similar manner as natural speech. These results demonstrate the existence of a multisensory speech-specific mode of perception.
Related Topics
Life Sciences Neuroscience Cognitive Neuroscience
Authors
, , , ,