...
首页> 外文期刊>Internet of Things Journal, IEEE >Adaptive Fusion and Category-Level Dictionary Learning Model for Multiview Human Action Recognition
【24h】

Adaptive Fusion and Category-Level Dictionary Learning Model for Multiview Human Action Recognition

机译:多维人类行动识别的自适应融合与类别级词典学习模型

获取原文
获取原文并翻译 | 示例

摘要

Human actions are often captured by multiple cameras (or sensors) to overcome the significant variations in viewpoints, background clutter, object speed, and motion patterns in video surveillance, and action recognition systems often benefit from fusing multiple types of cameras (sensors). Therefore, adaptive fusion of the information from multiple domains is mandatory for multiview human action recognition. Two widely applied fusion schemes are feature-level fusion and score-level fusion. We point out that limitations still exist and there is tremendous room for improvement, including the separate computation of feature fusion and action recognition, or the fixed weights for each action and each camera. However, previous fusion methods cannot accomplish them. In this paper, inspired by nature, the above limitations are addressed for multiview action recognition by developing a novel adaptive fusion and category-level dictionary learning model (abbreviated to AFCDL). It can jointly learn the adaptive weight for each camera and optimize the reconstruction of samples toward the action recognition task. To induce the dictionary learning and the reconstruction of query set (or test samples), the induced set for each category is built, and the corresponding induced regularization term is designed for the objective function. Extensive experiments on four public multiview action benchmarks show that AFCDL can significantly outperforms the state-of-the-art methods with 3% to 10% improvement in recognition accuracy.
机译:人类的行为通常由多个相机(或传感器)捕获,以克服视频监控中的视点,背景杂波,物体速度和运动模式的显着变化,并且动作识别系统通常受益于熔断多种类型的摄像机(传感器)。因此,来自多个域的信息的自适应融合是多视图人体行动识别的强制性。两个广泛应用的融合方案是特征级融合和得分级融合。我们指出,限制仍然存在,并且存在巨大的改进空间,包括单独计算特征融合和动作识别,或每个动作和每个相机的固定权重。但是,之前的融合方法无法完成它们。本文灵感来自自然的启发,通过开发新的自适应融合和类别级字典学习模型来解决以上的多视图动作识别(缩写为AFCDL)。它可以共同学习每个摄像机的自适应重量,并优化对动作识别任务的样本的重建。为了诱导字典学习和查询集的重建(或测试样本),构建了每个类别的感应集,并且相应的诱导正则化术语被设计用于目标函数。四个公共多维型动作基准的广泛实验表明,AFCDL能够显着优于最先进的方法,以提高识别准确性的3%至10%。

著录项

  • 来源
    《Internet of Things Journal, IEEE 》 |2019年第6期| 9280-9293| 共14页
  • 作者单位

    Qilu Univ Technol Shandong Acad Sci Shandong Comp Sci Ctr Shandong Artif Intelligence Inst Natl Supercomp C Jinan 250014 Shandong Peoples R China;

    Tianjin Univ Technol Minist Educ Key Lab Comp Vis & Syst Tianjin 300384 Peoples R China|Tianjin Univ Technol Tianjin Key Lab Intelligence Comp & Novel Softwar Tianjin 300384 Peoples R China;

    Tianjin Univ Technol Minist Educ Key Lab Comp Vis & Syst Tianjin 300384 Peoples R China|Tianjin Univ Technol Tianjin Key Lab Intelligence Comp & Novel Softwar Tianjin 300384 Peoples R China;

    Zhongnan Univ Econ & Law Sch Informat & Safety Engn Wuhan 430073 Hubei Peoples R China;

    Univ Texas San Antonio Dept Informat Syst & Cyber Secur San Antonio TX 78249 USA;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Adaptive fusion; category-level dictionary learning; improved dense trajectory (iDT); induced set; multiview action recognition;

    机译:自适应融合;类别级文字典学习;改进的致密轨迹(IDT);诱导集;多视图动作识别;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号