Speech listening entails neural encoding of invisible articulatory features

Autor: A. Pastore, A. Tomassini, I. Delis, E. Dolfini, L. Fadiga, A. D'Ausilio
Jazyk: angličtina
Rok vydání: 2022
Předmět:
Zdroj: NeuroImage, Vol 264, Iss , Pp 119724- (2022)
Druh dokumentu: article
ISSN: 1095-9572
DOI: 10.1016/j.neuroimage.2022.119724
Popis: Speech processing entails a complex interplay between bottom-up and top-down computations. The former is reflected in the neural entrainment to the quasi-rhythmic properties of speech acoustics while the latter is supposed to guide the selection of the most relevant input subspace. Top-down signals are believed to originate mainly from motor regions, yet similar activities have been shown to tune attentional cycles also for simpler, non-speech stimuli. Here we examined whether, during speech listening, the brain reconstructs articulatory patterns associated to speech production. We measured electroencephalographic (EEG) data while participants listened to sentences during the production of which articulatory kinematics of lips, jaws and tongue were also recorded (via Electro-Magnetic Articulography, EMA). We captured the patterns of articulatory coordination through Principal Component Analysis (PCA) and used Partial Information Decomposition (PID) to identify whether the speech envelope and each of the kinematic components provided unique, synergistic and/or redundant information regarding the EEG signals. Interestingly, tongue movements contain both unique as well as synergistic information with the envelope that are encoded in the listener's brain activity. This demonstrates that during speech listening the brain retrieves highly specific and unique motor information that is never accessible through vision, thus leveraging audio-motor maps that arise most likely from the acquisition of speech production during development.
Databáze: Directory of Open Access Journals