首页> 外文OA文献 >The contribution of visual information to the perception of speech in noise with and without informative temporal fine structure
【2h】

The contribution of visual information to the perception of speech in noise with and without informative temporal fine structure

机译:视觉信息对噪声中语音感知的贡献

摘要

Understanding what is said in demanding listening situations is assisted greatly by looking at the face of a talker. Previous studies have observed that normal-hearing listeners can benefit from this visual information when a talker’s voice is presented in background noise. These benefits have also been observed in quiet listening conditions in cochlear-implant users, whose device does not convey the informative temporal fine structure cues in speech, and when normal-hearing individuals listen to speech processed to remove these informative temporal fine structure cues. The current study (1) characterised the benefits of visual information when listening in background noise; and (2) used sine-wave vocoding to compare the size of the visual benefit when speech is presented with or without informative temporal fine structure. The accuracy with which normal-hearing individuals reported words in spoken sentences was assessed across three experiments. The availability of visual information and informative temporal fine structure cues was varied within and across the experiments. The results showed that visual benefit was observed using open- and closed-set tests of speech perception. The size of the benefit increased when informative temporal fine structure cues were removed. This finding suggests that visual information may play an important role in the ability of cochlear-implant users to understand speech in many everyday situations. Models of audio-visual integration were able to account for the additional benefit of visual information when speech was degraded and suggested that auditory and visual information was being integrated in a similar way in all conditions. The modelling results were consistent with the notion that audio-visual benefit is derived from the optimal combination of auditory and visual sensory cues.
机译:通过看着说话者的脸,可以大大帮助理解在苛刻的聆听情况下所说的话。以前的研究已经观察到,当说话者的声音以背景噪音呈现时,正常听力的听众可以从这些视觉信息中受益。这些益处也已在植入人工耳蜗的用户安静的聆听条件下观察到,他们的设备不能在语音中传达信息性的时间精细结构提示,并且当正常听力的人听取语音以去除这些信息性的时间精细结构提示时也是如此。当前的研究(1)描述了在听背景噪音时视觉信息的好处; (2)使用正弦波声码技术比较有或没有有益的时间精细结构的语音呈现时视觉效果的大小。在三个实验中评估了听力正常的人报告口语单词的准确性。在实验过程中和整个实验过程中,视觉信息和有用的时间精细结构提示的可用性均有所不同。结果表明,使用开放式和封闭式语音感知测试可观察到视觉益处。删除有益的时间精细结构提示后,收益的大小会增加。这一发现表明,视觉信息可能在许多日常情况下在人工耳蜗使用者理解语音的能力中发挥重要作用。当语音质量下降时,视听整合模型可以解决视觉信息的其他好处,并建议在所有情况下以相似的方式整合听觉和视觉信息。建模结果与以下概念一致:视听利益源于听觉和视觉感觉线索的最佳组合。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号