Skip to Main content Skip to Navigation
Journal articles

Bimodal speech: early suppressive visual effects in human auditory cortex.

Abstract : While everyone has experienced that seeing lip movements may improve speech perception, little is known about the neural mechanisms by which audiovisual speech information is combined. Event-related potentials (ERPs) were recorded while subjects performed an auditory recognition task among four different natural syllables randomly presented in the auditory (A), visual (V) or congruent bimodal (AV) condition. We found that: (i) bimodal syllables were identified more rapidly than auditory alone stimuli; (ii) this behavioural facilitation was associated with cross-modal [AV-(A+V)] ERP effects around 120-190 ms latency, expressed mainly as a decrease of unimodal N1 generator activities in the auditory cortex. This finding provides evidence for suppressive, speech-specific audiovisual integration mechanisms, which are likely to be related to the dominance of the auditory modality for speech perception. Furthermore, the latency of the effect indicates that integration operates at pre-representational stages of stimulus analysis, probably via feedback projections from visual and/or polymodal areas.
Document type :
Journal articles
Complete list of metadata

Cited literature [56 references]  Display  Hide  Download
Contributor : Julien Besle Connect in order to contact the contributor
Submitted on : Thursday, June 1, 2006 - 5:22:35 PM
Last modification on : Thursday, November 21, 2019 - 2:08:40 AM
Long-term archiving on: : Monday, April 5, 2010 - 10:09:52 PM




Julien Besle, Alexandra Fort, Claude Delpuech, Marie-Hélène Giard. Bimodal speech: early suppressive visual effects in human auditory cortex.. European Journal of Neuroscience, Wiley, 2004, 20 (8), pp.2225-34. ⟨10.1111/j.1460-9568.2004.03670.x⟩. ⟨inserm-00077887⟩



Record views


Files downloads