首页> 美国卫生研究院文献>Sensors (Basel Switzerland) >Deep Ego-Motion Classifiers for Compound Eye Cameras
【2h】

Deep Ego-Motion Classifiers for Compound Eye Cameras

机译:适用于复眼相机的深度自我运动分类器

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Compound eyes, also known as insect eyes, have a unique structure. They have a hemispheric surface, and a lot of single eyes are deployed regularly on the surface. Thanks to this unique form, using the compound images has several advantages, such as a large field of view (FOV) with low aberrations. We can exploit these benefits in high-level vision applications, such as object recognition, or semantic segmentation for a moving robot, by emulating the compound images that describe the captured scenes from compound eye cameras. In this paper, to the best of our knowledge, we propose the first convolutional neural network (CNN)-based ego-motion classification algorithm designed for the compound eye structure. To achieve this, we introduce a voting-based approach that fully utilizes one of the unique features of compound images, specifically, the compound images consist of a lot of single eye images. The proposed method classifies a number of local motions by CNN, and these local classifications which represent the motions of each single eye image, are aggregated to the final classification by a voting procedure. For the experiments, we collected a new dataset for compound eye camera ego-motion classification which contains scenes of the inside and outside of a certain building. The samples of the proposed dataset consist of two consequent emulated compound images and the corresponding ego-motion class. The experimental results show that the proposed method has achieved the classification accuracy of 85.0%, which is superior compared to the baselines on the proposed dataset. Also, the proposed model is light-weight compared to the conventional CNN-based image recognition algorithms such as AlexNet, ResNet50, and MobileNetV2.
机译:复眼,也称为昆虫眼,具有独特的结构。它们具有半球形表面,并且许多单眼规则地部署在该表面上。由于这种独特的形式,使用复合图像具有多个优点,例如具有低像差的大视场(FOV)。通过模拟描述从复眼摄像机捕获的场景的复合图像,我们可以在高级视觉应用中利用这些好处,例如对象识别或移动机器人的语义分割。在本文中,就我们所知,我们提出了第一种基于卷积神经网络(CNN)的自我运动分类算法,专为复眼结构设计。为此,我们引入了一种基于投票的方法,该方法充分利用了复合图像的独特功能之一,具体地说,复合图像由许多单眼图像组成。所提出的方法通过CNN对许多局部运动进行分类,并且这些代表每个单眼图像运动的局部分类通过投票程序汇总到最终分类中。对于实验,我们收集了用于复眼相机自我运动分类的新数据集,其中包含特定建筑物内部和外部的场景。所提出的数据集的样本包括两个随后的模拟复合图像和相应的自我运动类。实验结果表明,所提出的方法已达到85.0%的分类精度,优于所提出的数据集上的基线。而且,与传统的基于CNN的图像识别算法(例如AlexNet,ResNet50和MobileNetV2)相比,所提出的模型重量轻。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号