首页> 美国卫生研究院文献>Frontiers in Human Neuroscience >Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level
【2h】

Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level

机译:自上而下的神经水平视听语音知觉预测。

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。
获取外文期刊封面目录资料

摘要

During speech perception, listeners rely on multimodal input and make use of both auditory and visual information. When presented with speech, for example syllables, the differences in brain responses to distinct stimuli are not, however, caused merely by the acoustic or visual features of the stimuli. The congruency of the auditory and visual information and the familiarity of a syllable, that is, whether it appears in the listener’s native language or not, also modulates brain responses. We investigated how the congruency and familiarity of the presented stimuli affect brain responses to audio-visual (AV) speech in 12 adult Finnish native speakers and 12 adult Chinese native speakers. They watched videos of a Chinese speaker pronouncing syllables (/pa/, /pha/, /ta/, /tha/, /fa/) during a magnetoencephalography (MEG) measurement where only /pa/ and /ta/ were part of Finnish phonology while all the stimuli were part of Chinese phonology. The stimuli were presented in audio-visual (congruent or incongruent), audio only, or visual only conditions. The brain responses were examined in five time-windows: 75–125, 150–200, 200–300, 300–400, and 400–600 ms. We found significant differences for the congruency comparison in the fourth time-window (300–400 ms) in both sensor and source level analysis. Larger responses were observed for the incongruent stimuli than for the congruent stimuli. For the familiarity comparisons no significant differences were found. The results are in line with earlier studies reporting on the modulation of brain responses for audio-visual congruency around 250–500 ms. This suggests a much stronger process for the general detection of a mismatch between predictions based on lip movements and the auditory signal than for the top-down modulation of brain responses based on phonological information.
机译:在语音感知过程中,听众依靠多模式输入并利用听觉和视觉信息。但是,当以语音(例如音节)进行演讲时,大脑对不同刺激的反应差异并非仅由刺激的听觉或视觉特征引起。听觉和视觉信息的一致性以及音节的熟悉程度(即是否以听众的母语出现)也会调节大脑的反应。我们调查了所呈现的刺激的一致性和熟悉性如何影响12位芬兰成年人和12位中国成年人的大脑对视听(AV)语音的反应。他们在磁脑电图(MEG)测量中观看了讲汉语发音的音节(/ pa /,/ pha /,/ ta /,/ tha /,/ fa /)的视频,其中只有/ pa /和/ ta /是芬兰语的一部分语音学,而所有的刺激都是中国语音学的一部分。刺激以视听(一致或不一致),仅音频或仅视觉条件呈现。在五个时间窗口中检查了大脑的反应:75-125、150-200、200-300、300-400和400-600毫秒。我们在传感器和源电平分析的第四个时间窗口(300-400 ms)中发现一致性比较的显着差异。观察到,与全等刺激相比,非全等刺激的反应更大。对于熟悉程度的比较,没有发现显着差异。该结果与早期的研究报告一致,后者报道了250-500 ms左右的视听一致性的大脑反应调制。对于基于嘴唇运动和听觉信号的预测之间的不匹配的一般检测,这表明比基于语音信息的自上而下的大脑响应调制要强得多。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号