...
首页> 外文期刊>Affective Computing, IEEE Transactions on >Prediction-Based Audiovisual Fusion for Classification of Non-Linguistic Vocalisations
【24h】

Prediction-Based Audiovisual Fusion for Classification of Non-Linguistic Vocalisations

机译:基于预测的视听融合,用于非语言发声分类

获取原文
获取原文并翻译 | 示例

摘要

Prediction plays a key role in recent computational models of the brain and it has been suggested that the brain constantly makes multisensory spatiotemporal predictions. Inspired by these findings we tackle the problem of audiovisual fusion from a new perspective based on prediction. We train predictive models which model the spatiotemporal relationship between audio and visual features by learning the audio-to-visual and visual-to-audio feature mapping for each class. Similarly, we train predictive models which model the time evolution of audio and visual features by learning the past-to-future feature mapping for each class. In classification, all the class-specific regression models produce a prediction of the expected audio/visual features and their prediction errors are combined for each class. The set of class-specific regressors which best describes the audiovisual feature relationship, i.e., results in the lowest prediction error, is chosen to label the input frame. We perform cross-database experiments, using the AMI, SAL, and MAHNOB databases, in order to classify laughter and speech and subject-independent experiments on the AVIC database in order to classify laughter, hesitation and consent. In virtually all cases prediction-based audiovisual fusion consistently outperforms the two most commonly used fusion approaches, decision-level and feature-level fusion.
机译:预测在最近的大脑计算模型中起着关键作用,并且有人建议大脑不断做出多感觉的时空预测。受到这些发现的启发,我们从基于预测的新角度解决了视听融合问题。我们通过学习每个类别的视听和视听音频特征映射来训练预测模型,该模型对视听特征之间的时空关系建模。同样,我们通过学习每个类的过去到将来的特征映射来训练预测模型,这些模型对音频和视觉特征的时间演变进行建模。在分类中,所有特定于类别的回归模型都会对预期的音频/视频特征进行预测,并针对每个类别组合其预测误差。选择最能描述视听特征关系(即,导致最低的预测误差)的特定类回归器的集合来标记输入帧。我们使用AMI,SAL和MAHNOB数据库执行跨数据库实验,以便对笑声和语音进行分类,并在中航工业数据库上进行与受试者无关的实验,以便对笑声,犹豫和同意进行分类。在几乎所有情况下,基于预测的视听融合始终优于两种最常用的融合方法:决策级和特征级融合。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号