首页> 外文会议> >Semantic segmentation in egocentric video frames with deep learning for recognition of activities of daily living
【24h】

Semantic segmentation in egocentric video frames with deep learning for recognition of activities of daily living

机译:深度学习中的自我监视视频帧中的语义细分,以识别日常生活活动

获取原文
获取外文期刊封面目录资料

摘要

The analysis of videos for the recognition of Instrumental Activities of Daily Living (IADL) through the detection ofobjects and the context analysis, applied for the evaluation of patient’s capacity with Alzheimer's disease and age relateddementia, has recently gained a lot of interest. The incorporation of human perception in the recognition tasks, search,detection and visual content understanding has become one of the main tools for the development of systems andtechnologies that support the performance of people in their daily life activities. In this paper we propose a model ofautomatic segmentation of the saliency region where the objects of interest are found in egocentric video using fullyconvolutional networks (FCN). The segmentation is performed with the information regarding to human perception,obtaining a better segmentation at pixel level. This segmentation involves objects of interest and the salient region inegocentric videos, providing precise information to detection systems and automatic indexing of objects in video, wherethese systems have improved their performance in the recognition of IADL. To measure models segmentation performanceof the salient region, we benchmark two databases; first, Georgia-Tech-Egocentric-Activity database and second, our owndatabase. Results show that the method achieves a significantly better performance in the precision of the semanticsegmentation of the region where the objects of interest are located, compared with GBVS (Graph-Based Visual Saliency)method.
机译:通过检测分析了日常生活(IADL)仪器活动的视频 对象和上下文分析,申请评估患者容量与阿尔茨海默病和年龄相关的能力 痴呆症,最近获得了很多兴趣。在识别任务中纳入人类感知,搜索, 检测和视觉内容了解已成为系统和系统开发的主要工具之一 支持人们在日常生活活动中表现的技术。在本文中,我们提出了一种模型 自动分割显着区域,其中利用的自动媒体视频 卷积网络(FCN)。通过关于人类感知的信息进行分割, 在像素级别获得更好的分割。该分割涉及兴趣对象和突出区域 EGENENTRIC视频,为检测系统提供精确的信息和视频中对象的自动索引 这些系统在识别IADL时提高了它们的性能。测量模型分段性能 突出区域,我们基准两个数据库;一,乔治亚州科技 - Egentric-Activity数据库和第二,我们自己 数据库。结果表明,该方法在语义的精度下实现了明显更好的性能 与GBV(基于图形的视觉显着性)相比,感兴趣对象所在的区域的分割 方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号