首页> 外文会议>International conference on digital image processing >A New Method of 3D Scene Recognition from Still Images
【24h】

A New Method of 3D Scene Recognition from Still Images

机译:从静止图像识别3D场景的新方法

获取原文

摘要

Most methods of monocular visual three dimensional (3D) scene recognition involve supervised machine learning. However, these methods often rely on prior knowledge. Specifically, they learn the image scene as part of a training dataset. For this reason, when the sampling equipment or scene is changed, monocular visual 3D scene recognition may fail. To cope with this problem, a new method of unsupervised learning for monocular visual 3D scene recognition is here proposed. First, the image is made using superpixel segmentation based on the CIELAB color space values L, a, and b and on the coordinate values x and y of pixels, forming a superpixel image with a specific density. Second, a spectral clustering algorithm based on the superpixels' color characteristics and neighboring relationships was used to reduce the dimensions of the superpixel image. Third, the fuzzy distribution density functions representing sky, ground, and facade are multiplied with the segment pixels, where the expectations of these segments are obtained. A preliminary classification of sky, ground, and facade is generated in this way. Fourth, the most accurate classification images of sky, ground, and facade were extracted through the tier-1 wavelet sampling and Manhattan direction feature. Finally, a depth perception map is generated based on the pinhole imaging model and the linear perspective information of ground surface. Here, 400 images of Make3D Image data from the Cornell University website were used to test the algorithm. The experimental results showed that this unsupervised learning method provides a more effective monocular visual 3D scene recognition model than other methods.
机译:单眼视觉三维(3D)场景识别的大多数方法都涉及有监督的机器学习。但是,这些方法通常依赖于先验知识。具体来说,他们将图像场景作为训练数据集的一部分进行学习。因此,当更改采样设备或场景时,单目视觉3D场景识别可能会失败。为了解决这个问题,在此提出了一种用于单眼视觉3D场景识别的无监督学习新方法。首先,基于CIELAB颜色空间值L,a和b以及像素的坐标值x和y使用超像素分割来制作图像,从而形成具有特定密度的超像素图像。其次,基于超像素的颜色特征和相邻关系的光谱聚类算法被用于减小超像素图像的尺寸。第三,将代表天空,地面和立面的模糊分布密度函数与分段像素相乘,从而获得这些分段的期望值。以这种方式生成天空,地面和立面的初步分类。第四,通过1级小波采样和曼哈顿方向特征提取了最准确的天空,地面和立面分类图像。最后,基于针孔成像模型和地表的线性透视信息生成深度感知图。在这里,使用了康奈尔大学网站上的400张Make3D图像数据图像来测试该算法。实验结果表明,这种无监督学习方法比其他方法提供了更有效的单眼视觉3D场景识别模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号