首页> 外文期刊>Selected Topics in Applied Earth Observations and Remote Sensing, IEEE Journal of >Densely Based Multi-Scale and Multi-Modal Fully Convolutional Networks for High-Resolution Remote-Sensing Image Semantic Segmentation
【24h】

Densely Based Multi-Scale and Multi-Modal Fully Convolutional Networks for High-Resolution Remote-Sensing Image Semantic Segmentation

机译:基于密集的多尺度多模态全卷积网络用于高分辨率遥感图像语义分割

获取原文
获取原文并翻译 | 示例
           

摘要

Automatic and accurate semantic segmentation from high-resolution remote-sensing images plays an important role in the field of aerial images analysis. The task of dense semantic segmentation requires that semantic labels be assigned to each pixel in the image. Recently, convolutional neural networks (CNNs) have proven to be powerful tools for image classification, and they have been adopted in the remote-sensing community. But many limitations still exist when modern CNN architectures are directly applied to remote-sensing images, such as gradient explosion when the depth of the network increases, over-fitting with limited labeled remote-sensing data, and special differences between remote-sensing images and natural images. In this paper, we present a novel architecture that combines the thought of dense connection and fully convolutional networks, referred as DFCN, to automatically provide fine-grained semantic segmentation maps. In addition, we improve DFCN with multi-scale filters to widen the network and to increase the richness and diversity of extracted information, making the network more powerful and expressive than the naive convolution layer. Furthermore, we investigate a multi-modal network that incorporates digital surface models (DSMs) into a DFCN structure, and then we propose dual-path densely convolutional networks where the encoder consists of two paths that, respectively, extract features from spectral data and DSMs data and then fuse them. Finally, through conducting comprehensive experimental evaluations on two remote sensing benchmark datasets, we test our proposed models and compare them with other deep networks. The results demonstrate the effectiveness of proposed approaches; they can achieve competitive performance compared with the current state-of-the-art methods.
机译:从高分辨率遥感影像中进行自动,准确的语义分割在航空影像分析领域起着重要的作用。密集语义分割的任务要求将语义标签分配给图像中的每个像素。最近,卷积神经网络(CNN)已被证明是用于图像分类的强大工具,并且已在遥感领域被采用。但是,当现代CNN架构直接应用于遥感影像时,仍然存在许多局限性,例如网络深度增加时的梯度爆炸,对有限的带标记遥感数据的过度拟合,以及遥感影像与自然图像。在本文中,我们提出了一种新颖的架构,该架构结合了密集连接的思想和完全卷积的网络(称为DFCN),可以自动提供细粒度的语义分割图。此外,我们使用多尺度过滤器改进了DFCN,以扩大网络并增加提取信息的丰富性和多样性,从而使网络比朴素的卷积层更强大和更具表现力。此外,我们研究了将数字表面模型(DSM)合并到DFCN结构中的多模态网络,然后提出了双路径密集卷积网络,其中编码器由两条路径组成,分别从光谱数据和DSM中提取特征数据,然后将它们融合。最后,通过对两个遥感基准数据集进行全面的实验评估,我们测试了我们提出的模型并将其与其他深度网络进行了比较。结果证明了所提出方法的有效性;与目前最先进的方法相比,它们可以实现竞争优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号