...
首页> 外文期刊>PLoS Computational Biology >A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech
【24h】

A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech

机译:因果推理模型说明了麦格克效应和其他不一致的视听语音的感知

获取原文
           

摘要

Audiovisual speech integration combines information from auditory speech (talker’s voice) and visual speech (talker’s mouth movements) to improve perceptual accuracy. However, if the auditory and visual speech emanate from different talkers, integration decreases accuracy. Therefore, a key step in audiovisual speech perception is deciding whether auditory and visual speech have the same source, a process known as causal inference. A well-known illusion, the McGurk Effect, consists of incongruent audiovisual syllables, such as auditory “ba” + visual “ga” (AbaVga), that are integrated to produce a fused percept (“da”). This illusion raises two fundamental questions: first, given the incongruence between the auditory and visual syllables in the McGurk stimulus, why are they integrated; and second, why does the McGurk effect not occur for other, very similar syllables (e.g., AgaVba). We describe a simplified model of causal inference in multisensory speech perception (CIMS) that predicts the perception of arbitrary combinations of auditory and visual speech. We applied this model to behavioral data collected from 60 subjects perceiving both McGurk and non-McGurk incongruent speech stimuli. The CIMS model successfully predicted both the audiovisual integration observed for McGurk stimuli and the lack of integration observed for non-McGurk stimuli. An identical model without causal inference failed to accurately predict perception for either form of incongruent speech. The CIMS model uses causal inference to provide a computational framework for studying how the brain performs one of its most important tasks, integrating auditory and visual speech cues to allow us to communicate with others.
机译:视听语音集成结合了听觉语音(说话者的语音)和视觉语音(说话者的嘴巴动作)中的信息,以提高感知准确性。但是,如果听觉和视觉语音来自不同的讲话者,则整合会降低准确性。因此,视听语音感知中的关键步骤是确定听觉和视觉语音是否具有相同的来源,这一过程称为因果推理。 McGurk效应是一种众所周知的错觉,由听觉“ ba” +视觉“ ga”(AbaVga)等不协调的视听音节组成,这些音节被整合以产生融合感(da)。这种错觉提出了两个基本问题:首先,鉴于麦古尔克刺激中听觉和视觉音节之间的不一致,为什么它们被整合了?其次,为什么其他非常相似的音节(例如AgaVba)不会出现McGurk效应。我们描述了一种简化的多感官语音感知(CIMS)因果推理模型,该模型可预测听觉和视觉语音的任意组合。我们将此模型应用于从60名接受了McGurk和非McGurk非一致语音刺激的受试者中收集的行为数据。 CIMS模型成功地预测了对McGurk刺激观察到的视听整合以及对非McGurk刺激观察不到的整合。没有因果推理的相同模型无法准确预测任何形式的不一致语音的感知。 CIMS模型使用因果推理来提供计算框架,以研究大脑如何执行其最重要的任务之一,整合听觉和视觉语音提示以使我们能够与他人进行交流。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号