首页> 外文期刊>Sensors Journal, IEEE >Visual Object Recognition and Pose Estimation Based on a Deep Semantic Segmentation Network
【24h】

Visual Object Recognition and Pose Estimation Based on a Deep Semantic Segmentation Network

机译:基于深度语义分割网络的视觉目标识别与姿态估计

获取原文
获取原文并翻译 | 示例
       

摘要

In recent years, deep learning-based object recognition algorithms become emerging in robotic vision applications. This paper addresses the design of a novel deep learning-based visual object recognition and pose estimation system for a robot manipulator to handle random object picking tasks. The proposed visual control system consists of a visual perception module, an object pose estimation module, a data argumentation module, and a robot manipulator controller. The visual perception module combines deep convolution neural networks (CNNs) and a fully connected conditional random field layer to realize an image semantic segmentation function, which can provide stable and accurate object classification results in cluttered environments. The object pose estimation module implements a model-based pose estimation method to estimate the 3D pose of the target for picking control. In addition, the proposed data argumentation module automatically generates training data for training the deep CNN. Experimental results show that the proposed scene segmentation method used in the data argumentation module reaches a high accuracy rate of 97.10% on average, which is higher than other state-of-the-art segment methods. Moreover, with the proposed data argumentation module, the visual perception module reaches an accuracy rate over than 80% and 72% in the case of detecting and recognizing one object and three objects, respectively. In addition, the proposed model-based pose estimation method provides accurate 3D pose estimation results. The average translation and rotation errors in the three axes are all smaller than 0.52 cm and 3.95 degrees, respectively. These advantages make the proposed visual control system suitable for applications of random object picking and manipulation.
机译:近年来,基于深度学习的对象识别算法在机器人视觉应用中崭露头角。本文介绍了一种新颖的基于深度学习的视觉对象识别和姿态估计系统的设计,该系统用于机器人操纵器来处理随机对象拾取任务。所提出的视觉控制系统包括视觉感知模块,对象姿态估计模块,数据论证模块和机器人操纵器控制器。视觉感知模块结合了深度卷积神经网络(CNN)和完全连接的条件随机场层来实现图像语义分割功能,可以在杂乱的环境中提供稳定,准确的对象分类结果。对象姿势估计模块实现基于模型的姿势估计方法以估计用于拾取控制的目标的3D姿势。另外,所提出的数据论证模块自动生成用于训练深度CNN的训练数据。实验结果表明,该算法在数据论证模块中使用的场景分割方法平均准确率达到97.10%,高于其他最新的分割方法。此外,通过提出的数据论证模块,在分别检测和识别一个物体和三个物体的情况下,视觉感知模块的准确率超过80%和72%。另外,所提出的基于模型的姿势估计方法提供了准确的3D姿势估计结果。三个轴上的平均平移和旋转误差均分别小于0.52 cm和3.95度。这些优点使所提出的视觉控制系统适合于随机物体拾取和操纵的应用。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号