首页> 外文期刊>Journal of electrical and computer engineering >Traffic Scene Depth Analysis Based on Depthwise Separable Convolutional Neural Network
【24h】

Traffic Scene Depth Analysis Based on Depthwise Separable Convolutional Neural Network

机译:基于深度可分离卷积神经网络的交通场景深度分析

获取原文
获取原文并翻译 | 示例
       

摘要

In order to obtain the distances between the surrounding objects and the vehicle in the traffic scene in front of the vehicle, a monocular visual depth estimation method based on the depthwise separable convolutional neural network is proposed in this study. First, features containing shallow depth information were extracted from the RGB images using the convolution layers and maximum pooling layers. Subsampling operations were also performed on these images. Subsequently, features containing advanced depth information were extracted using a block based on an ensemble of convolution layers and a block based on depth separable convolution layers. The output from all different blocks is combined afterwards. Finally, transposed convolution layers were used for upsampling the feature maps to the same size with the original RGB image. During the upsampling process, skip connections were used to merge the features containing shallow depth information that was obtained from the convolution operation through the depthwise separable convolution layers. The depthwise separable convolution layers can provide more accurate depth information features for estimating the monocular visual depth. At the same time, they require reduced computational cost and fewer parameter numbers while providing a similar level (or slightly better) computing performance. Integrating multiple simple convolutions into a block not only increases the overall depth of the neural network but also enables a more accurate extraction of the advanced features in the neural network. Combining the output from multiple blocks can prevent the loss of features containing important depth information. The testing results show that the depthwise separable convolutional neural network provides a superior performance than the other monocular visual depth estimation methods. Therefore, applying depthwise separable convolution layers in the neural network is a more effective and accurate approach for estimating the visual depth.
机译:为了在车辆前面的交通场景中获得周围物体和车辆之间的距离,在该研究中提出了一种基于深度可分离卷积神经网络的单眼视觉深度估计方法。首先,使用卷积层和最大池层从RGB图像中提取包含浅深度信息的特征。在这些图像上也执行了分支操作。随后,使用基于卷积层的集合和基于深度可分离卷积层的块来提取包含高级深度信息的特征。之后组合所有不同块的输出。最后,转换卷积层用于将特征映射到与原始RGB图像相同的大小。在上采样过程中,跳过连接用于合并包含从卷积操作通过深度可分离的卷积层获得的浅深度信息的特征。深度可分离的卷积层可以提供更精确的深度信息特征,用于估计单眼视觉深度。同时,它们需要减少计算成本和更少的参数编号,同时提供类似的级别(或稍好)计算性能。将多个简单卷积集成到块中不仅增加了神经网络的整体深度,而且还可以更准确地提取神经网络中的高级功能。组合来自多个块的输出可以防止包含重要深度信息的功能的丢失。测试结果表明,深度可分离的卷积神经网络提供比其他单眼视觉深度估计方法的优越性。因此,在神经网络中应用深度可分离的卷积层是一种更有效和准确的方法,用于估计视觉深度。

著录项

  • 来源
    《Journal of electrical and computer engineering》 |2019年第1期|9340129.1-9340129.10|共10页
  • 作者单位

    Zhejiang Univ Sci & Technol Sch Informat & Elect Engn Hangzhou 310023 Zhejiang Peoples R China;

    Zhejiang Univ Sci & Technol Sch Informat & Elect Engn Hangzhou 310023 Zhejiang Peoples R China|Zhejiang Univ Inst Informat & Commun Engn Hangzhou 310027 Zhejiang Peoples R China;

    Zhejiang Univ Sci & Technol Sch Informat & Elect Engn Hangzhou 310023 Zhejiang Peoples R China;

    Zhejiang Univ Sci & Technol Sch Informat & Elect Engn Hangzhou 310023 Zhejiang Peoples R China;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

  • 入库时间 2022-08-18 22:02:36

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号