首页> 外文期刊>Intelligent Service Robotics >Building 3D semantic maps for mobile robots using RGB-D camera
【24h】

Building 3D semantic maps for mobile robots using RGB-D camera

机译:使用RGB-D相机构建移动机器人的3D语义地图

获取原文
获取原文并翻译 | 示例
           

摘要

The wide availability of affordable RGB-D sensors changes the landscape of indoor scene analysis. Years of research on simultaneous localization and mapping (SLAM) have made it possible to merge multiple RGB-D images into a single point cloud and provide a 3D model for a complete indoor scene. However, these reconstructed models only have geometry information, not including semantic knowledge. The advancements in robot autonomy and capabilities for carrying out more complex tasks in unstructured environments can be greatly enhanced by endowing environment models with semantic knowledge. Towards this goal, we propose a novel approach to generate 3D semantic maps for an indoor scene. Our approach creates a 3D reconstructed map from a RGB-D image sequence firstly, then we jointly infer the semantic object category and structural class for each point of the global map. 12 object categories (e.g. walls, tables, chairs) and 4 structural classes (ground, structure, furniture and props) are labeled in the global map. In this way, we can totally understand both the object and structure information. In order to get semantic information, we compute semantic segmentation for each RGB-D image and merge the labeling results by a Dense Conditional Random Field. Different from previous techniques, we use temporal information and higher-order cliques to enforce the label consistency for each image labeling result. Our experiments demonstrate that temporal information and higher-order cliques are significant for the semantic mapping procedure and can improve the precision of the semantic mapping results.
机译:经济实惠的RGB-D传感器的广泛可用性改变了室内场景分析的景观。多年的同时定位和映射(SLAM)的研究使得可以将多个RGB-D图像合并到单点云中,并为完整的室内场景提供3D模型。然而,这些重建的模型仅具有几何信息,而不包括语义知识。通过具有语义知识的环境模型,可以大大提高机器人自主权和在非结构化环境中进行更复杂任务的能力。为了实现这一目标,我们提出了一种新颖的方法来为室内场景产生3D语义地图。我们的方法首先从RGB-D图像序列创建一个3D重建映射,然后我们共同推断全球地图每个点的语义对象类别和结构类。 12对象类别(例如墙壁,桌子,椅子)和4个结构类(地面,结构,家具和道具)在全球地图中标记。通过这种方式,我们可以完全理解对象和结构信息。为了获得语义信息,我们计算每个RGB-D图像的语义分割,并通过密集的条件随机字段合并标签结果。与以前的技术不同,我们使用时间信息和高阶批变来强制每个图像标记结果的标签一致性。我们的实验表明,时间信息和高阶派系对于语义映射过程很重要,并且可以提高语义映射结果的精度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号