Audiovisual speech entrainment in deafness
Audiovisual speech entrainment in deafness
Disciplines
Medical-Theoretical Sciences, Pharmacy (100%)
Keywords
-
Cochlear Implant,
Lipreading,
Deafness,
Entrainment
Communication requires listeners to correctly process the acoustic input delivered by a speaker. However, in many life situations the acoustic speech signal is accompanied and preceded by movements of the speaker, that are visually processed by the listener. Especially lip-movements have been shown to be beneficial for understanding speech, particularly in adverse listening situations, e.g. when trying to follow a conversation at a cocktail party. Also, many individuals with hearing loss depend on the ability to lip read. This ability has also been implicated to mediate rehabilitation of speech perception in individuals that become deaf later in development and receive a so-called cochlear implant (CI). This device circumvents the damaged inner-ear by directly stimulating the hearing nerve. However, especially for adults that have been used to "normal" acoustic input, the input via the CI appears very much distorted and the individual has to actively learn to assign meaning to the unusual sound impressions. Unfortunately, clinical outcomes of CI are tainted by extreme variability, leading to frustrating experiences and rejection of the device in some patients. Considering the costs and risks of the intervention, this variability is a serious problem. A common question to all of these raised issues is: How does the processing of lip movements improve auditory processing? We will capitalize on recent developments also in our lab, to track brain regions that faithfully follow speech-related acoustic and visual information at a rate closely resembling the important frequency of syllables. We will test how different levels of acoustic and visual noise impacts this relevant ability of the brain. Importantly, based on ongoing works we predict an involvement of visual brain regions in tracking the acoustic input also when missing in the stimulation. According to our view, this process is relevant when "translating" the visual input (i.e. lip movement) into a "format" that can be used by the auditory system. According to our project, this capability may be highly variable and partially driving the individual`s capability to lip read. Besides of testing normal hearing individuals, we will investigate how this process is changed in deaf individuals, who putatively rely significantly more on visual processing of lip movements. Finally, we will test how this neural tracking process changes following CI and how measures at an early stage could predict rehabilitation success. This project will use innovative approaches to provide important insights on processing of speech- related lip-movements and how these neural processes are affected by auditory deprivation. Our research could break new grounds in understanding how interindividual differences in neural tracking of lip movements (and its associated acoustic signal) could affect CI rehabilitation success, which is an urgent clinical issue.
When speaking, not only is an acoustic signal produced, but there are also accompanying visible lip movements. These lip movements are not a meaningless side effect, but can shape auditory processing in a significant way. This can be seen, for example, in the fact that seeing lip movements helps comprehension in noisy environments. This might be related to brain waves following lip movements. Brain activity also follows the acoustic speech signal, so-called speech tracking, even if the acoustic part of the speech was not presented at all, but only the lip movements without sound. This project investigates both mechanisms (tracking of the lips and the speech signal) with regard to some general details (e.g. in difficult listening situations) and in particular with regard to hearing difficulties up to deafness. For this purpose, we presented distorted speech material in a study and investigated speech tracking across different degrees of distortion (from well understandable to completely unintelligible). For normal-hearing subjects, we found that speech tracking is strongest for moderately distorted speech and decreases for easy-to-understand and incomprehensible speech. This speech distortion is an approximation of the hearing experience with cochlear implants (CI), so in another study we investigated speech tracking in postlingually deafened subjects (i.e. deafness after language acquisition) who received a CI. Recordings of brain activity after 6 and 12 months post implantation show that auditory speech processing normalises over the months, especially in the ear without a CI. In other studies in which the participant groups only saw lip movements, speech tracking was found in a younger participant group while it was reduced in an older group of subjects, which could be related to increasing hearing difficulties in the course of ageing. Prelingually deafened subjects (i.e. deafened before language acquisition) however showed stronger lip tracking. Since this speech tracking is mainly found in brain regions that are responsible for visual processing, we also investigated whether speech tracking is associated with so-called lip reading. We therefore also developed a test for lip reading skills in German, but no effects were found in this respect. This project has stimulated several new research directions focusing on the role of the visual system in language processing.
- Universität Salzburg - 100%
- Olivier Collignon, Université Catholique de Louvain - Belgium
- Joachim Gross, Universität Münster - Germany
- Jonas Obleser, Universität zu Lübeck - Germany
- Anne Keitel, University of Glasgow
Research Output
- 148 Citations
- 19 Publications
- 3 Datasets & models
- 1 Fundings
-
2021
Title Sound-location specific alpha power modulation in the visual cortex in absence of visual input DOI 10.5167/uzh-233443 Type Other Author Langer Link Publication -
2023
Title Brain areas associated with visual spatial attention display topographic organization during auditory spatial attention DOI 10.5167/uzh-233442 Type Other Author Gips Link Publication -
2024
Title Individual prediction tendencies do not generalize across modalities. DOI 10.1111/psyp.14435 Type Journal Article Author Schubert J Journal Psychophysiology -
2019
Title Degradation levels of continuous speech affect neural speech tracking and alpha power differently DOI 10.1101/615302 Type Preprint Author Hauswald A Pages 615302 Link Publication -
2023
Title Neural speech tracking shifts from the syllabic to the modulation rate of speech as intelligibility decreases. DOI 10.1111/psyp.14362 Type Journal Article Author Chen Yp Journal Psychophysiology -
2023
Title Individual prediction tendencies do not generalise across modalities DOI 10.1101/2023.02.02.526758 Type Preprint Author Schubert J -
2023
Title Speech intelligibility changes the temporal evolution of neural speech tracking. DOI 10.1016/j.neuroimage.2023.119894 Type Journal Article Author Chen Yp Journal NeuroImage Pages 119894 -
2022
Title Masking of the mouth area impairs reconstruction of acoustic speech features and higher-level segmentational features in the presence of a distractor speaker DOI 10.1016/j.neuroimage.2022.119044 Type Journal Article Author Haider C Journal NeuroImage Pages 119044 Link Publication -
2022
Title Cortical tracking of formant modulations derived from silently presented lip movements and its decline with age DOI 10.1093/cercor/bhab518 Type Journal Article Author Suess N Journal Cerebral Cortex Pages 4818-4833 Link Publication -
2022
Title Speech intelligibility changes the temporal evolution of neural speech tracking DOI 10.1101/2022.06.26.497639 Type Preprint Author Chen Y Pages 2022.06.26.497639 Link Publication -
2021
Title Sound-location specific alpha power modulation in the visual cortex in absence of visual input DOI 10.1101/2021.03.15.435371 Type Preprint Author Popov T Pages 2021.03.15.435371 Link Publication -
2021
Title Differential attention-dependent adjustment of frequency, power and phase in primary sensory and frontoparietal areas DOI 10.1016/j.cortex.2021.01.008 Type Journal Article Author Suess N Journal Cortex Pages 179-193 Link Publication -
2021
Title Cortical tracking of unheard formant modulations derived from silently presented lip movements and its decline with age DOI 10.1101/2021.04.13.439628 Type Preprint Author Suess N Pages 2021.04.13.439628 Link Publication -
2021
Title Occlusion of lip movements impairs reconstruction of acoustic speech features and higher-level segmentational features in the presence of a distractor speaker DOI 10.1101/2021.09.28.461909 Type Preprint Author Haider C Pages 2021.09.28.461909 Link Publication -
2019
Title Differential attention-dependent adjustment of frequency, power and phase in primary sensory and frontoparietal areas DOI 10.1101/697615 Type Preprint Author Suess N Pages 697615 Link Publication -
2020
Title Degradation levels of continuous speech affect neural speech tracking and alpha power differently DOI 10.1111/ejn.14912 Type Journal Article Author Hauswald A Journal European Journal of Neuroscience Pages 3288-3302 Link Publication -
2021
Title Neural speech tracking shifts from the syllabic to the modulation rate of speech as intelligibility decreases DOI 10.1101/2021.03.25.437033 Type Preprint Author Schmidt F Pages 2021.03.25.437033 Link Publication -
2022
Title Influence of linguistic properties and hearing impairment on visual speech perception skills in the German language DOI 10.1371/journal.pone.0275585 Type Journal Article Author Suess N Journal PLoS ONE Link Publication -
2022
Title Brain areas associated with visual spatial attention display topographic organization during auditory spatial attention DOI 10.1093/cercor/bhac285 Type Journal Article Author Popov T Journal Cerebral Cortex Pages 3478-3489 Link Publication
-
2021
Link
Title Influence of linguistic properties and hearing impairment on visual speech perception skills in the German language Type Database/Collection of data Public Access Link Link -
2021
Link
Title MEG data during silently presented lip movements and lip reading skills in participant group with a broad age-range Type Database/Collection of data Public Access Link Link -
2020
Link
Title MEG data during listening to degraded speech Type Database/Collection of data Public Access Link Link
-
2020
Title Impact of face masks on speech comprehension Type Other Start of Funding 2020 Funder Austrian Science Fund (FWF)