Auditory cortex encodes lipreading information through spatially distributed activity.

Autor: Karthik G; Department of Psychology, University of Michigan, Ann Arbor, MI 48109, USA., Cao CZ; Department of Psychology, University of Michigan, Ann Arbor, MI 48109, USA., Demidenko MI; Department of Psychology, University of Michigan, Ann Arbor, MI 48109, USA., Jahn A; Department of Psychology, University of Michigan, Ann Arbor, MI 48109, USA., Stacey WC; Department of Neurology, University of Michigan, Ann Arbor, MI 48109, USA., Wasade VS; Henry Ford Hospital, Detroit, MI 48202, USA; Department of Neurology, Wayne State University School of Medicine, Detroit, MI 48201, USA., Brang D; Department of Psychology, University of Michigan, Ann Arbor, MI 48109, USA. Electronic address: djbrang@umich.edu.
Jazyk: angličtina
Zdroj: Current biology : CB [Curr Biol] 2024 Sep 09; Vol. 34 (17), pp. 4021-4032.e5. Date of Electronic Publication: 2024 Aug 16.
DOI: 10.1016/j.cub.2024.07.073
Abstrakt: Watching a speaker's face improves speech perception accuracy. This benefit is enabled, in part, by implicit lipreading abilities present in the general population. While it is established that lipreading can alter the perception of a heard word, it is unknown how these visual signals are represented in the auditory system or how they interact with auditory speech representations. One influential, but untested, hypothesis is that visual speech modulates the population-coded representations of phonetic and phonemic features in the auditory system. This model is largely supported by data showing that silent lipreading evokes activity in the auditory cortex, but these activations could alternatively reflect general effects of arousal or attention or the encoding of non-linguistic features such as visual timing information. This gap limits our understanding of how vision supports speech perception. To test the hypothesis that the auditory system encodes visual speech information, we acquired functional magnetic resonance imaging (fMRI) data from healthy adults and intracranial recordings from electrodes implanted in patients with epilepsy during auditory and visual speech perception tasks. Across both datasets, linear classifiers successfully decoded the identity of silently lipread words using the spatial pattern of auditory cortex responses. Examining the time course of classification using intracranial recordings, lipread words were classified at earlier time points relative to heard words, suggesting a predictive mechanism for facilitating speech. These results support a model in which the auditory system combines the joint neural distributions evoked by heard and lipread words to generate a more precise estimate of what was said.
Competing Interests: Declaration of interests The authors declare no competing interests.
(Copyright © 2024 Elsevier Inc. All rights reserved.)
Databáze: MEDLINE