首页> 外文期刊>Computer vision and image understanding >You-Do, I-Learn: Egocentric unsupervised discovery of objects and their modes of interaction towards video-based guidance
【24h】

You-Do, I-Learn: Egocentric unsupervised discovery of objects and their modes of interaction towards video-based guidance

机译:您做,我学习:以自我为中心的无监督发现对象及其交互方式,以视频为指导

获取原文
获取原文并翻译 | 示例

摘要

This paper presents an unsupervised approach towards automatically extracting video-based guidance on object usage, from egocentric video and wearable gaze tracking, collected from multiple users while performing tasks. The approach (ⅰ) discovers task relevant objects, (ⅱ) builds a model for each, (ⅲ) distinguishes different ways in which each discovered object has been used and (ⅳ) discovers the dependencies between object interactions. The work investigates using appearance, position, motion and attention, and presents results using each and a combination of relevant features. Moreover, an online scalable approach is presented and is compared to offline results. The paper proposes a method for selecting a suitable video guide to be displayed to a novice user indicating how to use an object, purely triggered by the user's gaze. The potential assistive mode can also recommend an object to be used next based on the learnt sequence of object interactions. The approach was tested on a variety of daily tasks such as initialising a printer, preparing a coffee and setting up a gym machine.
机译:本文提出了一种无监督的方法,该方法可从以自我为中心的视频和可穿戴式凝视跟踪中自动提取基于视频的对象使用指导,这些对象是在执行任务时从多个用户收集的。方法(ⅰ)发现与任务相关的对象,(ⅱ)为每个对象建立一个模型,(ⅲ)区分使用每个发现对象的不同方式,(ⅳ)发现对象交互之间的依赖关系。该作品使用外观,位置,动作和注意力进行调查,并使用相关特征中的每一个及其组合来呈现结果。此外,提出了一种在线可扩展方法,并将其与离线结果进行了比较。本文提出了一种方法,用于选择要显示给新手用户的视频指南,指示如何使用对象,这完全是由用户的注视触发的。潜在的辅助模式还可以根据学习到的对象交互顺序推荐下一个要使用的对象。该方法已在各种日常任务上进行了测试,例如初始化打印机,准备咖啡和设置健身机。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号