Auditory speech detection in noise enhanced by lipreading
Autor: | Sumiko Takayanagi, Lynne E. Bernstein, Edward T. Auer |
---|---|
Rok vydání: | 2004 |
Předmět: |
Speechreading
Linguistics and Language Voice activity detection Visual perception Computer science Communication Speech recognition White noise Stimulus (physiology) Speech processing Language and Linguistics Computer Science Applications Modeling and Simulation Dynamic rectangle Detection theory Computer Vision and Pattern Recognition Software |
Zdroj: | Speech Communication. 44:5-18 |
ISSN: | 0167-6393 |
DOI: | 10.1016/j.specom.2004.10.011 |
Popis: | Audiovisual speech stimuli have been shown to produce a variety of perceptual phenomena. Enhanced detectability of acoustic speech in noise, when the talker can also be seen, is one of those phenomena. This study investigated whether this enhancement effect is specific to visual speech stimuli or can rely on more generic non-speech visual stimulus properties. Speech detection thresholds for an auditory /ba/ stimulus were obtained in a white noise masker. The auditory /ba/ was presented adaptively to obtain its 79.4% detection threshold under five conditions. In Experiment 1, the syllable was presented (1) auditory-only (AO) and (2) as audiovisual speech (AVS), using the original video recording. Three types of synthetic visual stimuli were also paired synchronously with the audio token: (3) A dynamic Lissajous (AVL) figure whose vertical extent was correlated with the acoustic speech envelope; (4) a dynamic rectangle (AVR) whose horizontal extent was correlated with the speech envelope; and (5) a static rectangle (AVSR) whose onset and offset were synchronous with the acoustic speech onset and offset. Ten adults with normal hearing and vision participated. The results, in terms of dB signal-to-noise ratio (SNR), were AVS |
Databáze: | OpenAIRE |
Externí odkaz: |