首页> 外文期刊>Trends in Hearing >Audiovisual Speech Recognition With a Cochlear Implant and Increased Perceptual and Cognitive Demands
【24h】

Audiovisual Speech Recognition With a Cochlear Implant and Increased Perceptual and Cognitive Demands

机译:视听演讲识别与耳蜗植入和增加的感知和认知需求

获取原文
           

摘要

Speech recognition in complex environments involves focusing on the most relevant speech signal while ignoring distractions. Difficulties can arise due to the incoming signal’s characteristics (e.g., accented pronunciation, background noise, distortion) or the listener’s characteristics (e.g., hearing loss, advancing age, cognitive abilities). Listeners who use cochlear implants (CIs) must overcome these difficulties while listening to an impoverished version of the signals available to listeners with normal hearing (NH). In the real world, listeners often attempt tasks concurrent with, but unrelated to, speech recognition. This study sought to reveal the effects of visual distraction and performing a simultaneous visual task on audiovisual speech recognition. Two groups, those with CIs and those with NH listening to vocoded speech, were presented videos of unaccented and accented talkers with and without visual distractions, and with a secondary task. It was hypothesized that, compared with those with NH, listeners with CIs would be less influenced by visual distraction or a secondary visual task because their prolonged reliance on visual cues to aid auditory perception improves the ability to suppress irrelevant information. Results showed that visual distractions alone did not significantly decrease speech recognition performance for either group, but adding a secondary task did. Speech recognition was significantly poorer for accented compared with unaccented speech, and this difference was greater for CI listeners. These results suggest that speech recognition performance is likely more dependent on incoming signal characteristics than a difference in adaptive strategies for managing distractions between those who listen with and without a CI.
机译:复杂环境中的语音识别涉及关注最相关的语音信号,同时忽略分心。由于传入信号的特征(例如,重音发音,背景噪声,失真)或听众的特征,可能出现困难(例如,听力损失,推进年龄,认知能力)。使用耳蜗植入物(CIS)的侦听器必须克服这些困难,同时收听具有正常听力(NH)的侦听器可用的贫困信号的贫困版本。在现实世界中,听众经常尝试与语音识别不相关的并发任务。本研究试图揭示视觉分散注意力的影响,并对视听语音识别进行同时视觉任务。有两组与CIS的人和有NH听声音演讲的人,被提出了无人组成的谈话者的视频,并没有视觉分心,并具有二次任务。它被假设,与NH的人相比,随着视觉分心或次要视觉任务的影响,具有CIS的听众会影响,因为它们的长期依赖视觉提示,以帮助听觉感知提高抑制无关信息的能力。结果表明,单独的视觉分心并没有显着降低任一组的语音识别性能,而是添加二级任务。与未集中的言论相比,语音识别显着较差,而CI听众的差异更大。这些结果表明,语音识别性能可能更依赖于传入的信号特性,而不是管理在倾听和没有CI的人之间分心的自适应策略的差异。

著录项

获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号