...
首页> 外文期刊>Affective Computing, IEEE Transactions on >Identifying Human Behaviors Using Synchronized Audio-Visual Cues
【24h】

Identifying Human Behaviors Using Synchronized Audio-Visual Cues

机译:使用同步视听提示识别人类行为

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

In this paper, a human behavior recognition method using multimodal features is presented. We focus on modeling individual and social behaviors of a subject (e.g., friendly/aggressive or hugging/kissing behaviors) with a hidden conditional random field (HCRF) in a supervised framework. Each video is represented by a vector of spatio-temporal visual features (STIP, head orientation and proxemic features) along with audio features (MFCCs). We propose a feature pruning method for removing irrelevant and redundant features based on the spatio-temporal neighborhood of each feature in a video sequence. The proposed framework assumes that human movements are highly correlated with sound emissions. For this reason, canonical correlation analysis (CCA) is employed to find correlation between the audio and video features prior to fusion. The experimental results, performed in two human behavior recognition datasets including political speeches and human interactions from TV shows, attest the advantages of the proposed method compared with several baseline and alternative human behavior recognition methods.
机译:本文提出了一种利用多峰特征的人类行为识别方法。我们专注于在监督框架中使用隐藏的条件随机字段(HCRF)对主题的个体和社交行为进行建模(例如,友好/攻击性或拥抱/亲吻行为)。每个视频都由时空视觉特征(STIP,头部朝向和近距离特征)以及音频特征(MFCC)组成。我们提出了一种基于视频序列中每个特征的时空邻域特征的删除不相关和冗余特征的特征修剪方法。拟议的框架假设人类的运动与声音的发射高度相关。因此,在融合之前,采用规范相关分析(CCA)来找到音频和视频特征之间的相关性。在两个人类行为识别数据集中进行的实验结果,包括政治演讲和电视节目中的人类互动,证明了与几种基准和其他人类行为识别方法相比,该方法的优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号