Audio-visual speech timing sensitivity is enhanced in cluttered conditions.

Events encoded in separate sensory modalities, such as audition and vision, can seem to be synchronous across a relatively broad range of physical timing differences. This may suggest that the precision of audio-visual timing judgments is inherently poor. Here we show that this is not necessarily tr...

Descripción completa

Guardado en:
Detalles Bibliográficos
Autores principales: Warrick Roseboom, Shin'ya Nishida, Waka Fujisaki, Derek H Arnold
Formato: article
Lenguaje:EN
Publicado: Public Library of Science (PLoS) 2011
Materias:
R
Q
Acceso en línea:https://doaj.org/article/02a8e834d7114fc680a93862cd1ae2de
Etiquetas: Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!
id oai:doaj.org-article:02a8e834d7114fc680a93862cd1ae2de
record_format dspace
spelling oai:doaj.org-article:02a8e834d7114fc680a93862cd1ae2de2021-11-18T06:56:08ZAudio-visual speech timing sensitivity is enhanced in cluttered conditions.1932-620310.1371/journal.pone.0018309https://doaj.org/article/02a8e834d7114fc680a93862cd1ae2de2011-04-01T00:00:00Zhttps://www.ncbi.nlm.nih.gov/pmc/articles/pmid/21494684/pdf/?tool=EBIhttps://doaj.org/toc/1932-6203Events encoded in separate sensory modalities, such as audition and vision, can seem to be synchronous across a relatively broad range of physical timing differences. This may suggest that the precision of audio-visual timing judgments is inherently poor. Here we show that this is not necessarily true. We contrast timing sensitivity for isolated streams of audio and visual speech, and for streams of audio and visual speech accompanied by additional, temporally offset, visual speech streams. We find that the precision with which synchronous streams of audio and visual speech are identified is enhanced by the presence of additional streams of asynchronous visual speech. Our data suggest that timing perception is shaped by selective grouping processes, which can result in enhanced precision in temporally cluttered environments. The imprecision suggested by previous studies might therefore be a consequence of examining isolated pairs of audio and visual events. We argue that when an isolated pair of cross-modal events is presented, they tend to group perceptually and to seem synchronous as a consequence. We have revealed greater precision by providing multiple visual signals, possibly allowing a single auditory speech stream to group selectively with the most synchronous visual candidate. The grouping processes we have identified might be important in daily life, such as when we attempt to follow a conversation in a crowded room.Warrick RoseboomShin'ya NishidaWaka FujisakiDerek H ArnoldPublic Library of Science (PLoS)articleMedicineRScienceQENPLoS ONE, Vol 6, Iss 4, p e18309 (2011)
institution DOAJ
collection DOAJ
language EN
topic Medicine
R
Science
Q
spellingShingle Medicine
R
Science
Q
Warrick Roseboom
Shin'ya Nishida
Waka Fujisaki
Derek H Arnold
Audio-visual speech timing sensitivity is enhanced in cluttered conditions.
description Events encoded in separate sensory modalities, such as audition and vision, can seem to be synchronous across a relatively broad range of physical timing differences. This may suggest that the precision of audio-visual timing judgments is inherently poor. Here we show that this is not necessarily true. We contrast timing sensitivity for isolated streams of audio and visual speech, and for streams of audio and visual speech accompanied by additional, temporally offset, visual speech streams. We find that the precision with which synchronous streams of audio and visual speech are identified is enhanced by the presence of additional streams of asynchronous visual speech. Our data suggest that timing perception is shaped by selective grouping processes, which can result in enhanced precision in temporally cluttered environments. The imprecision suggested by previous studies might therefore be a consequence of examining isolated pairs of audio and visual events. We argue that when an isolated pair of cross-modal events is presented, they tend to group perceptually and to seem synchronous as a consequence. We have revealed greater precision by providing multiple visual signals, possibly allowing a single auditory speech stream to group selectively with the most synchronous visual candidate. The grouping processes we have identified might be important in daily life, such as when we attempt to follow a conversation in a crowded room.
format article
author Warrick Roseboom
Shin'ya Nishida
Waka Fujisaki
Derek H Arnold
author_facet Warrick Roseboom
Shin'ya Nishida
Waka Fujisaki
Derek H Arnold
author_sort Warrick Roseboom
title Audio-visual speech timing sensitivity is enhanced in cluttered conditions.
title_short Audio-visual speech timing sensitivity is enhanced in cluttered conditions.
title_full Audio-visual speech timing sensitivity is enhanced in cluttered conditions.
title_fullStr Audio-visual speech timing sensitivity is enhanced in cluttered conditions.
title_full_unstemmed Audio-visual speech timing sensitivity is enhanced in cluttered conditions.
title_sort audio-visual speech timing sensitivity is enhanced in cluttered conditions.
publisher Public Library of Science (PLoS)
publishDate 2011
url https://doaj.org/article/02a8e834d7114fc680a93862cd1ae2de
work_keys_str_mv AT warrickroseboom audiovisualspeechtimingsensitivityisenhancedinclutteredconditions
AT shinyanishida audiovisualspeechtimingsensitivityisenhancedinclutteredconditions
AT wakafujisaki audiovisualspeechtimingsensitivityisenhancedinclutteredconditions
AT derekharnold audiovisualspeechtimingsensitivityisenhancedinclutteredconditions
_version_ 1718424146131550208