...
首页> 外文期刊>eLife journal >Contributions of local speech encoding and functional connectivity to audio-visual speech perception
【24h】

Contributions of local speech encoding and functional connectivity to audio-visual speech perception

机译:本地语音编码和功能连接对视听语音感知的贡献

获取原文

摘要

When listening to someone in a noisy environment, such as a cocktail party, we can understand the speaker more easily if we can also see his or her face. Movements of the lips and tongue convey additional information that helps the listener’s brain separate out syllables, words and sentences. However, exactly where in the brain this effect occurs and how it works remain unclear. To find out, Giordano et al. scanned the brains of healthy volunteers as they watched clips of people speaking. The clarity of the speech varied between clips. Furthermore, in some of the clips the lip movements of the speaker corresponded to the speech in question, whereas in others the lip movements were nonsense babble. As expected, the volunteers performed better on a word recognition task when the speech was clear and when the lips movements agreed with the spoken dialogue. Watching the video clips stimulated rhythmic activity in multiple regions of the volunteers’ brains, including areas that process sound and areas that plan movements. Speech is itself rhythmic, and the volunteers’ brain activity synchronized with the rhythms of the speech they were listening to. Seeing the speaker’s face increased this degree of synchrony. However, it also made it easier for sound-processing regions within the listeners’ brains to transfer information to one other. Notably, only the latter effect predicted improved performance on the word recognition task. This suggests that seeing a person’s face makes it easier to understand his or her speech by boosting communication between brain regions, rather than through effects on individual areas. Further work is required to determine where and how the brain encodes lip movements and speech sounds. The next challenge will be to identify where these two sets of information interact, and how the brain merges them together to generate the impression of specific words.
机译:在嘈杂的环境中(例如鸡尾酒会)听某人说话时,如果我们还能看到说话者的脸,那么我们可以更轻松地理解说话者。嘴唇和舌头的运动传达了更多信息,有助于听众的大脑将音节,单词和句子分开。但是,尚不清楚这种作用在大脑中的何处发生以及如何发挥作用。为了找到答案,佐丹奴等人。看着健康的志愿者观看人们讲话的片段时,他们扫描了他们的大脑。语音的清晰度在片段之间有所不同。此外,在某些剪辑中,说话者的唇部运动与所讨论的语音相对应,而在其他剪辑中,唇部运动是胡说八道。正如预期的那样,当语音清晰,嘴唇动作与语音对话一致时,志愿者在单词识别任务上的表现会更好。观看视频片段刺激了志愿者大脑多个区域的节律活动,包括处理声音的区域和计划运动的区域。语音本身具有节奏感,志愿者的大脑活动与他们正在听的语音节奏保持同步。看到说话者的脸可以增加这种同步度。但是,这也使听众大脑中的声音处理区域更易于彼此传递信息。值得注意的是,只有后一种效果才能预测单词识别任务的性能提高。这表明,看到一个人的脸可以促进大脑区域之间的交流,而不是通过对单个区域的影响,从而更容易理解他或她的言语。需要进一步的工作来确定大脑在哪里以及如何编码嘴唇运动和语音。下一个挑战将是确定这两套信息在哪里相互作用,以及大脑如何将它们融合在一起以产生特定单词的印象。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号