首页> 外文会议>International Conference on Multimodal Interfaces >Modeling Naturalistic Affective States Via Facial, Vocal, and Bodily Expressions Recognition
【24h】

Modeling Naturalistic Affective States Via Facial, Vocal, and Bodily Expressions Recognition

机译:通过面部,声乐和身体表达识别建模自然主义情感状态

获取原文

摘要

Affective and human-centered computing have attracted a lot of attention during the past years, mainly due to the abundance of devices and environments able to exploit multimodal input from the part of the users and adapt their functionality to their preferences or individual habits. In the quest to receive feedback from the users in an unobtrusive manner, the combination of facial and hand gestures with prosody information allows us to infer the users' emotional state, relying on the best performing modality in cases where one modality suffers from noise or bad sensing conditions. In this paper, we describe a multi-cue, dynamic approach to detect emotion in naturalistic video sequences. Contrary to strictly controlled recording conditions of audiovisual material, the proposed approach focuses on sequences taken from nearly real world situations. Recognition is performed via a 'Simple Recurrent Network' which lends itself well to modeling dynamic events in both user's facial expressions and speech. Moreover this approach differs from existing work in that it models user expressivity using a dimensional representation of activation and valence, instead of detecting discrete 'universal emotions', which are scarce in everyday human-machine interaction. The algorithm is deployed on an audiovisual database which was recorded simulating human-human discourse and, therefore, contains less extreme expressivity and subtle variations of a number of emotion labels.
机译:情感和以人为本的计算在过去几年中引起了很多关注,主要是由于能够从用户部分利用多模式输入的设备和环境的丰富,并适应他们的偏好或个人习惯。在寻求以不引人注目的方式接收用户的反馈意见中,面部和手势与韵律信息的组合使我们能够推断用户的情绪状态,依赖于一个模态遭受噪音或坏的情况下最好的表现形态传感条件。在本文中,我们描述了一种多重线,动态方法来检测自然视频序列中的情绪。与受音像材料的严格控制的记录条件相反,所提出的方法侧重于从几乎现实世界情况所采取的序列。通过“简单复制网络”执行识别,其对用户的面部表情和语音中的动态事件进行了良好的建模。此外,这种方法与现有的工作不同,因为它模拟了使用激活和价值的尺寸表示的用户表征,而不是检测在日常人机交互中稀缺的离散的“普遍情绪”。该算法部署在录制的视听数据库上,该数据库被记录在模拟人类话语中,因此,含有较少的极端表达性和许多情感标签的微妙变化。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号