首页> 外文期刊>Journal of Imaging >ECRU: An Encoder-Decoder Based Convolution Neural Network (CNN) for Road-Scene Understanding
【24h】

ECRU: An Encoder-Decoder Based Convolution Neural Network (CNN) for Road-Scene Understanding

机译:ECRU:一种基于编码器-解码器的卷积神经网络(CNN),用于路演场景理解

获取原文
       

摘要

This research presents the idea of a novel fully-Convolutional Neural Network (CNN)-based model for probabilistic pixel-wise segmentation, titled Encoder-decoder-based CNN for Road-Scene Understanding (ECRU). Lately, scene understanding has become an evolving research area, and semantic segmentation is the most recent method for visual recognition. Among vision-based smart systems, the driving assistance system turns out to be a much preferred research topic. The proposed model is an encoder-decoder that performs pixel-wise class predictions. The encoder network is composed of a VGG-19 layer model, while the decoder network uses 16 upsampling and deconvolution units. The encoder of the network has a very flexible architecture that can be altered and trained for any size and resolution of images. The decoder network upsamples and maps the low-resolution encoder’s features. Consequently, there is a substantial reduction in the trainable parameters, as the network recycles the encoder’s pooling indices for pixel-wise classification and segmentation. The proposed model is intended to offer a simplified CNN model with less overhead and higher performance. The network is trained and tested on the famous road scenes dataset CamVid and offers outstanding outcomes in comparison to similar early approaches like FCN and VGG16 in terms of performance vs. trainable parameters.
机译:这项研究提出了一种基于新型全卷积神经网络(CNN)的概率像素分割模型的想法,该模型的标题为基于编码器-解码器的CNN,用于场景理解(ECRU)。最近,场景理解已成为一个不断发展的研究领域,而语义分割是用于视觉识别的最新方法。在基于视觉的智能系统中,驾驶辅助系统被证明是首选的研究主题。所提出的模型是执行像素级类别预测的编码器-解码器。编码器网络由VGG-19层模型组成,而解码器网络使用16个上采样和反卷积单元。网络的编码器具有非常灵活的体系结构,可以针对任何大小和分辨率的图像进行更改和培训。解码器网络对低分辨率编码器的功能进行升采样和映射。因此,随着网络回收编码器的池索引以进行像素分类和分割,可训练参数将大大减少。提出的模型旨在提供一种具有更少开销和更高性能的简化CNN模型。该网络在著名的道路场景数据集CamVid上进行了训练和测试,与类似的早期方法(如FCN和VGG16)相比,在性能和可训练参数方面提供了出色的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号