首页> 外文期刊>IEEE transactions on automation science and engineering >Incremental Learning of Object Models From Natural Human–Robot Interactions
【24h】

Incremental Learning of Object Models From Natural Human–Robot Interactions

机译:自然人机器人交互的对象模型的增量学习

获取原文
获取原文并翻译 | 示例
       

摘要

In order to perform complex tasks in realistic human environments, robots need to be able to learn new concepts in the wild, incrementally, and through their interactions with humans. This article presents an end-to-end pipeline to learn object models incrementally during the human-robot interaction (HRI). The pipeline we propose consists of three parts: 1) recognizing the interaction type; 2) detecting the object that the interaction is targeting; and 3) learning incrementally the models from data recorded by the robot sensors. Our main contributions lie in the target object detection, guided by the recognized interaction, and in the incremental object learning. The novelty of our approach is the focus on natural, heterogeneous, and multimodal HRIs to incrementally learn new object models. Throughout the article, we highlight the main challenges associated with this problem, such as high degree of occlusion and clutter, domain change, low-resolution data, and interaction ambiguity. This article shows the benefits of using multiview approaches and combining visual and language features, and our experimental results outperform standard baselines. Note to Practitioners-This article was motivated by challenges in recognition tasks for dynamic and varying scenarios. Our approach learns to recognize new user interactions and objects. To do so, we use multimodal data from the user-robot interaction; visual data are used to learn the objects and speech is used to learn the label and help with the interaction-type recognition. We use state-of-the-art deep learning (DL) models to segment the user and the objects in the scene. Our algorithm for incremental learning is based on a classic incremental clustering approach. The pipeline we propose works with all sensors mounted on the robot, so it allows mobility on the system. This article uses the data recorded from a Baxter robot, which enables the use of the manipulation arms in future steps, but it would work with any robot that can have the same sensors mounted. The sensors used are two RGB-D cameras and a microphone. The pipeline currently has high computational requirements to run the two DL-based steps. We have tested it with a desktop computer, including a GTX 1060 and 32 GB of RAM.
机译:为了在现实的人类环境中执行复杂的任务,机器人需要能够在野外,逐步地和与人类的互动中学习新概念。本文介绍了端到端管道,用于在人机交互(HRI)期间逐步学习对象模型。我们提出的管道由三部分组成:1)识别互动类型; 2)检测相互作用靶向的对象; 3)学习从机器人传感器记录的数据逐步学习。我们的主要贡献位于目标对象检测,由公认的交互引导,并在增量对象学习中。我们的方法的新颖性是对天然,异构和多模式HRI的重点是逐步学习新的对象模型。在整个文章中,我们突出了与这个问题相关的主要挑战,例如高度的遮挡和杂波,域变化,低分辨率数据以及相互作用的歧义。本文显示了使用多视图方法和结合视觉和语言特征的好处,以及我们的实验结果优于标准基线。从业者的注意事项 - 本文受到动态和不同情景的承认任务的挑战。我们的方法学会识别新的用户交互和对象。为此,我们使用来自用户机器人交互的多模式数据;可视化数据用于了解对象,语音用于学习标签并帮助交互类型识别。我们使用最先进的深度学习(DL)模型来分割用户和场景中的对象。我们的增量学习算法基于经典增量聚类方法。我们提出的管道与安装在机器人上的所有传感器一起使用,因此它允许系统上的移动性。本文使用从Baxter Robot记录的数据,这使得可以在未来的步骤中使用操纵臂,但它将适用于任何可以安装相同传感器的机器人。使用的传感器是两个RGB-D相机和麦克风。管道目前具有高计算要求来运行基于两个DL的步骤。我们已经使用台式计算机进行了测试,包括GTX 1060和32 GB的RAM。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号