Article ID | Journal | Published Year | Pages | File Type |
---|---|---|---|---|
10458109 | Cognition | 2005 | 10 Pages |
Abstract
In face-to-face conversation speech is perceived by ear and eye. We studied the prerequisites of audio-visual speech perception by using perceptually ambiguous sine wave replicas of natural speech as auditory stimuli. When the subjects were not aware that the auditory stimuli were speech, they showed only negligible integration of auditory and visual stimuli. When the same subjects learned to perceive the same auditory stimuli as speech, they integrated the auditory and visual stimuli in a similar manner as natural speech. These results demonstrate the existence of a multisensory speech-specific mode of perception.
Related Topics
Life Sciences
Neuroscience
Cognitive Neuroscience
Authors
Jyrki Tuomainen, Tobias S. Andersen, Kaisa Tiippana, Mikko Sams,