...
首页> 外文期刊>Quality Control, Transactions >Transfer Learning for Humanoid Robot Appearance-Based Localization in a Visual Map
【24h】

Transfer Learning for Humanoid Robot Appearance-Based Localization in a Visual Map

机译:在视觉地图中转移人形机器人出现的本地化的学习

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Autonomous robot visual navigation is a fundamental locomotion task based on extracting relevant features from images taken from the surrounded environment to control an independent displacement. In the navigation, the use of a known visual map helps obtain an accurate localization, but in the absence of this map, a guided or free exploration pathway must be executed to obtain the images sequence representing the visual map. This paper presents an appearance-based localization method based on a visual map and an end-to-end Convolutional Neural Network (CNN). The CNN is initialized via transfer learning (trained using the ImageNet dataset), evaluating four state-of-the-art CNN architectures: VGG16, ResNet50, InceptionV3, and Xception. A typical pipeline for transfer learning includes changing the last layer to adapt the number of neurons according to the number of custom classes. In this work, the dense layers after the convolutional and pooling layers were substituted by a Global Average Pooling (GAP) layer, which is parameter-free. Additionally, an $L_{2}$ -norm constraint was added to the GAP layer feature descriptors, restricting the features from lying on a fixed radius hypersphere. These different pre-trained configurations were analyzed and compared using two visual maps found in the CIMAT-NAO datasets consisting of 187 and 94 images, respectively. For evaluating the localization tasks, a set of 278 and 94 images were available for each visual map, respectively. The numerical results proved that by integrating the $L_{2}$ -norm constraint in the training pipeline, the appearance-based localization performance is boosted. Specifically, the pre-trained VGG16 and Xception networks achieved the best localization results, reaching a top-3 accuracy of 90.70% and 93.62% for each dataset, respectively, overcoming the referenced approaches based on hand-crafted feature extractors.
机译:自主机器人视觉导航是基于从周围环境中拍摄的图像中提取的相关特征来控制独立位移的基本运动任务。在导航中,使用已知的视觉图有助于获得准确的本地化,而是在没有此地图的情况下,必须执行引导或自由探索途径以获得表示视觉图的图像序列。本文介绍了一种基于视觉地图和端到端卷积神经网络(CNN)的基于外观的本地化方法。 CNN通过传输学习初始化(使用ImageNet DataSet培训),评估四个最先进的CNN架构:VGG16,Resnet50,Inceptionv3和七。用于转移学习的典型管道包括改变最后一层以根据自定义类的数量来调整神​​经元的数量。在这项工作中,卷积和池池层后的致密层被全局平均水平池(间隙)层代替,这是无参数的。此外,<内联公式XMLNS:MML =“http://www.w3.org/1998/math/mathml”xmlns:xlink =“http://www.w3.org/1999/xlink”> $ l_ {2} $ -norm约束被添加到间隙层特征描述符,限制了符合固定的半径超短的功能。通过分别由187和94个图像组成的CIMAT-NAO数据集中发现的两个视觉映射来分析和比较这些不同的预先训练的配置。为了评估本地化任务,分别可用于每个可视图的278和94个图像。证明了数值结果证明,通过集成<内联公式XMLNS:MML =“http://www.w3.org/1998/math/mathml”xmlns:xlink =“http://www.w3.org/1999/ XLink“> $ L_ {2} $ -Norm在训练管道中的约束,提高了基于外观的本地化性能。具体而言,预先训练的VGG16和Xcepion网络达到了最佳的本地化结果,分别为每个数据集达到90.70%和93.62%的前3个精度,克服了基于手工制作的特征提取器的引用方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号