首页> 外文期刊>Neurocomputing >Deep appearance and motion learning for egocentric activity recognition
【24h】

Deep appearance and motion learning for egocentric activity recognition

机译:深度外观和动作学习,以自我为中心的活动识别

获取原文
获取原文并翻译 | 示例
           

摘要

Egocentric activity recognition has recently generated great popularity in computer vision due to its widespread applications in egocentric video analysis. However, it poses new challenges comparing to the conventional third-person activity recognition tasks, which are caused by significant body shaking, varied lengths, and poor recoding quality, etc. To handle these challenges, in this paper, we propose deep appearance and motion learning (DAML) for egocentric activity recognition, which leverages the great strength of deep learning networks in feature learning. In contrast to hand- crafted visual features or pre-trained convolutional neural network (CNN) features with limited generality to new egocentric videos, the proposed DAML is built on the deep autoencoder (DAE), and directly extracts appearance and motion feature, the main cue of activities, from egocentric videos. The DAML takes advantages of the great effectiveness and efficiency of the DAE in unsupervised feature learning, which provides a new representation learning framework of egocentric videos. The learned appearance and motion features by the DAML are seamlessly fused to accomplish a rich informative egocentric activity representation which can be readily fed into any supervised learning models for activity recognition. Experimental results on two challenging benchmark datasets show that the DAML achieves high performance on both short- and long-term egocentric activity recognition tasks, which is comparable to or even better than the state-of-the-art counterparts. (C) 2017 Elsevier B.V. All rights reserved.
机译:以自我为中心的活动识别由于在以自我为中心的视频分析中的广泛应用,最近在计算机视觉中获得了极大的普及。但是,与传统的第三人称活动识别任务相比,它带来了新的挑战,这是由明显的身体晃动,长度变化和记录质量差等引起的。为应对这些挑战,在本文中,我们提出了深层的外观和动作学习(DAML)以自我为中心的活动识别,它利用深度学习网络在功能学习中的强大优势。与手工制作的视觉特征或预训练的卷积神经网络(CNN)特征对新的以自我为中心的视频具有有限的通用性相比,拟议的DAML建立在深度自动编码器(DAE)的基础上,并直接提取外观和运动特征。来自自我中心视频的活动提示。 DAML在无人监督的特征学习中利用了DAE的巨大有效性和效率,从而为以自我为中心的视频提供了新的表示学习框架。 DAML将学习到的外观和动作功能无缝融合,以完成丰富的以自我为中心的活动表示,可以很容易地将其输入到任何受监督的学习模型中以进行活动识别。在两个具有挑战性的基准数据集上的实验结果表明,DAML在短期和长期的以自我为中心的活动识别任务上均实现了卓越的性能,这与最新的同类活动相当甚至更好。 (C)2017 Elsevier B.V.保留所有权利。

著录项

  • 来源
    《Neurocomputing》 |2018年第31期|438-447|共10页
  • 作者单位

    Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Sichuan, Peoples R China;

    Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Sichuan, Peoples R China;

    Columbia Univ, Sch Engn & Appl Sci, New York, NY 10027 USA;

    Univ Western Ontario, Digital Imaging Grp, London, ON N6A 4V2, Canada;

    Univ Trento, Dept Informat Engn & Comp Sci, I-38100 Trento, Italy;

    Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Sichuan, Peoples R China;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Multiple feature learning; Deep learning; Autoencoder; Egocentric video; Activity recognition;

    机译:多特征学习;深度学习;自动编码器;以自我为中心的视频;活动识别;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号