首页> 外文期刊>The International journal of robotics research >One-shot visual appearance learning for mobile manipulation
【24h】

One-shot visual appearance learning for mobile manipulation

机译:一键式视觉外观学习,可进行移动操作

获取原文
获取原文并翻译 | 示例
           

摘要

We describe a vision-based algorithm that enables a robot to robustly detect specific objects in a scene following an initial segmentation hint from a human user. The novelty lies in the ability to 'reacquire' objects over extended spatial and temporal excursions within challenging environments based upon a single training example. The primary difficulty lies in achieving an effective reacquisition capability that is robust to the effects of local clutter, lighting variation, and object relocation. We overcome these challenges through an adaptive detection algorithm that automatically generates multiple-view appearance models far each object online. As the robot navigates within the environment and the object is detected from different viewpoints, the one-shot learner opportunistically and automatically incorporates additional observations into each model. In order to overcome the effects of 'drift' common to adaptive learners, the algorithm imposes simple requirements on the geometric consistency of candidate observations. Motivating our reacquisition strategy is our work developing a mobile manipulator that interprets and autonomously performs commands conveyed by a human user. The ability to detect specific objects and reconstitute the user s segmentation hints enables the robot to be situationally aware. This situational awareness enables rich command and control mechanisms and affords natural interaction. We demonstrate one such capability that allows the human to give the robot a 'guided tour' of named objects within an outdoor environment and, hours later, to direct the robot to manipulate those objects by name using spoken instructions. We implemented our appearance-based detection strategy on our robotic manipulator as it operated over multiple days in different outdoor environments. We evaluate the algorithm's performance under challenging conditions that include scene clutter, lighting and viewpoint variation, object ambiguity, and object relocation. The results demonstrate a reacquisition capability that is effective in real-world settings.
机译:我们描述了一种基于视觉的算法,该算法使机器人能够根据人类用户的初始分割提示来稳健地检测场景中的特定对象。新颖之处在于能够根据单个训练示例在挑战性环境中的扩展的空间和时间偏移中“重新获取”对象。主要困难在于实现有效的重新捕获功能,该功能对局部杂波,照明变化和对象重定位的影响具有鲁棒性。我们通过自适应检测算法克服了这些挑战,该算法自动在线生成远距离每个对象的多视图外观模型。当机器人在环境中导航并且从不同的视角检测到对象时,单发学习者就会有机会地自动将其他观察结果合并到每个模型中。为了克服适应性学习者常见的“漂移”效应,该算法对候选观测值的几何一致性提出了简单的要求。激励我们的重新获得策略的工作是我们开发一种移动机械手,该机械手解释并自主执行人类用户传达的命令。检测特定对象并重新构造用户的细分提示的能力使机器人可以根据情况进行感知。这种态势感知可以实现丰富的命令和控制机制,并提供自然的交互作用。我们演示了一种这样的功能,它允许人类在室外环境中对机器人进行命名对象的“导览”,并在数小时后指示机器人使用语音指令按名称操作这些对象。由于我们的机器人在不同的室外环境中运行了数天,因此我们在机器人上实施了基于外观的检测策略。我们在具有挑战性的条件下评估算法的性能,这些条件包括场景混乱,照明和视点变化,对象模糊性以及对象重定位。结果证明了在现实环境中有效的重新采集功能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号