Home > Research > Publications & Outputs > Reassessing the Benefits of Audiovisual Integra...

Electronic data

Links

Text available via DOI:

View graph of relations

Reassessing the Benefits of Audiovisual Integration to Speech Perception and Intelligibility

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Published

Standard

Reassessing the Benefits of Audiovisual Integration to Speech Perception and Intelligibility. / O'Hanlon, Brandon; Plack, Christopher J.; Nuttall, Helen E.
In: Journal of Speech, Language, and Hearing Research, Vol. 68, No. 1, 02.01.2025, p. 26-39.

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Harvard

APA

Vancouver

O'Hanlon B, Plack CJ, Nuttall HE. Reassessing the Benefits of Audiovisual Integration to Speech Perception and Intelligibility. Journal of Speech, Language, and Hearing Research. 2025 Jan 2;68(1):26-39. Epub 2024 Dec 2. doi: 10.1044/2024_jslhr-24-00162

Author

Bibtex

@article{b32b1855d9e74691877ee86b4ac5f291,
title = "Reassessing the Benefits of Audiovisual Integration to Speech Perception and Intelligibility",
abstract = "Purpose: In difficult listening conditions, the visual system assists with speech perception through lipreading. Stimulus onset asynchrony (SOA) is used to investigate the interaction between the two modalities in speech perception. Previous estimates of audiovisual benefit and SOA integration period differ widely. A limitation of previous research is a lack of consideration of visemes—categories of phonemes defined by similar lip movements when produced by a speaker—to ensure that selected phonemes are visually distinct. This study aimed to reassess the benefits of audiovisual lipreading to speech perception when different viseme categories are selected as stimuli and presented in noise. The study also aimed to investigate the effects of SOA on these stimuli. Method: Sixty participants were tested online and presented with audio-only and audiovisual stimuli containing the speaker's lip movements. The speech was presented either with or without noise and had six different SOAs (0, 200, 216.6, 233.3, 250, and 266.6 ms). Participants discriminated between speech syllables with button presses. Results: The benefit of visual information was weaker than that in previous studies. There was a significant increase in reaction times as SOA was introduced, but there were no significant effects of SOA on accuracy. Furthermore, exploratory analyses suggest that the effect was not equal across viseme categories: “Ba” was more difficult to recognize than “ka” in noise. Conclusion: In summary, the findings suggest that the contributions of audiovisual integration to speech processing are weaker when considering visemes but are not sufficient to identify a full integration period. Supplemental Material: https://doi.org/10.23641/asha.27641064",
author = "Brandon O'Hanlon and Plack, {Christopher J.} and Nuttall, {Helen E.}",
year = "2025",
month = jan,
day = "2",
doi = "10.1044/2024_jslhr-24-00162",
language = "English",
volume = "68",
pages = "26--39",
journal = "Journal of Speech, Language, and Hearing Research",
issn = "1092-4388",
publisher = "American Speech-Language-Hearing Association (ASHA)",
number = "1",

}

RIS

TY - JOUR

T1 - Reassessing the Benefits of Audiovisual Integration to Speech Perception and Intelligibility

AU - O'Hanlon, Brandon

AU - Plack, Christopher J.

AU - Nuttall, Helen E.

PY - 2025/1/2

Y1 - 2025/1/2

N2 - Purpose: In difficult listening conditions, the visual system assists with speech perception through lipreading. Stimulus onset asynchrony (SOA) is used to investigate the interaction between the two modalities in speech perception. Previous estimates of audiovisual benefit and SOA integration period differ widely. A limitation of previous research is a lack of consideration of visemes—categories of phonemes defined by similar lip movements when produced by a speaker—to ensure that selected phonemes are visually distinct. This study aimed to reassess the benefits of audiovisual lipreading to speech perception when different viseme categories are selected as stimuli and presented in noise. The study also aimed to investigate the effects of SOA on these stimuli. Method: Sixty participants were tested online and presented with audio-only and audiovisual stimuli containing the speaker's lip movements. The speech was presented either with or without noise and had six different SOAs (0, 200, 216.6, 233.3, 250, and 266.6 ms). Participants discriminated between speech syllables with button presses. Results: The benefit of visual information was weaker than that in previous studies. There was a significant increase in reaction times as SOA was introduced, but there were no significant effects of SOA on accuracy. Furthermore, exploratory analyses suggest that the effect was not equal across viseme categories: “Ba” was more difficult to recognize than “ka” in noise. Conclusion: In summary, the findings suggest that the contributions of audiovisual integration to speech processing are weaker when considering visemes but are not sufficient to identify a full integration period. Supplemental Material: https://doi.org/10.23641/asha.27641064

AB - Purpose: In difficult listening conditions, the visual system assists with speech perception through lipreading. Stimulus onset asynchrony (SOA) is used to investigate the interaction between the two modalities in speech perception. Previous estimates of audiovisual benefit and SOA integration period differ widely. A limitation of previous research is a lack of consideration of visemes—categories of phonemes defined by similar lip movements when produced by a speaker—to ensure that selected phonemes are visually distinct. This study aimed to reassess the benefits of audiovisual lipreading to speech perception when different viseme categories are selected as stimuli and presented in noise. The study also aimed to investigate the effects of SOA on these stimuli. Method: Sixty participants were tested online and presented with audio-only and audiovisual stimuli containing the speaker's lip movements. The speech was presented either with or without noise and had six different SOAs (0, 200, 216.6, 233.3, 250, and 266.6 ms). Participants discriminated between speech syllables with button presses. Results: The benefit of visual information was weaker than that in previous studies. There was a significant increase in reaction times as SOA was introduced, but there were no significant effects of SOA on accuracy. Furthermore, exploratory analyses suggest that the effect was not equal across viseme categories: “Ba” was more difficult to recognize than “ka” in noise. Conclusion: In summary, the findings suggest that the contributions of audiovisual integration to speech processing are weaker when considering visemes but are not sufficient to identify a full integration period. Supplemental Material: https://doi.org/10.23641/asha.27641064

U2 - 10.1044/2024_jslhr-24-00162

DO - 10.1044/2024_jslhr-24-00162

M3 - Journal article

VL - 68

SP - 26

EP - 39

JO - Journal of Speech, Language, and Hearing Research

JF - Journal of Speech, Language, and Hearing Research

SN - 1092-4388

IS - 1

ER -