首页> 外文会议>International Conference on Artificial Neural Networks >Unsharp Masking Layer: Injecting Prior Knowledge in Convolutional Networks for Image Classification
【24h】

Unsharp Masking Layer: Injecting Prior Knowledge in Convolutional Networks for Image Classification

机译:锐化的遮罩层:在卷积网络中注入先验知识以进行图像分类

获取原文

摘要

Image enhancement refers to the enrichment of certain image features such as edges, boundaries, or contrast. The main objective is to process the original image so that the overall performance of visualization, classification and segmentation tasks is considerably improved. Traditional techniques require manual fine-tuning of the parameters to control enhancement behavior. To date, recent Convolutional Neural Network (CNN) approaches frequently employ the aforementioned techniques as an enriched pre-processing step. In this work, we present the first intrinsic CNN pre-processing layer based on the well-known unsharp masking algorithm. The proposed layer injects prior knowledge about how to enhance the image, by adding high frequency information to the input, to subsequently emphasize meaningful image features. The layer optimizes the unsharp masking parameters during model training, without any manual intervention. We evaluate the network performance and impact on two applications: CIFAR100 image classification, and the PlantCLEF identification challenge. Results obtained show a significant improvement over popular CNNs, yielding 9.49% and 2.42% for PlantCLEF and general-purpose CIFAR100, respectively. The design of an unsharp enhancement layer plainly boosts the accuracy with negligible performance cost on simple CNN models, as prior knowledge is directly injected to improve its robustness.
机译:图像增强是指丰富某些图像特征,例如边缘,边界或对比度。主要目的是处理原始图像,以便显着提高可视化,分类和分割任务的整体性能。传统技术需要手动微调参数以控制增强行为。迄今为止,最近的卷积神经网络(CNN)方法经常将上述技术用作丰富的预处理步骤。在这项工作中,我们基于众所周知的unsharp masking算法提出了第一个固有的CNN预处理层。所提议的层通过向输入中添加高频信息来注入有关如何增强图像的先验知识,以随后强调有意义的图像特征。该层在模型训练期间优化了不清晰的遮罩参数,而无需任何手动干预。我们评估了网络性能及其对两种应用的影响:CIFAR100图像分类和PlantCLEF识别挑战。获得的结果表明,与流行的CNN相比,PlantCLEF和通用CIFAR100的收率分别提高了9.49%和2.42%。不清晰的增强层的设计在简单的CNN模型上以可忽略的性能成本明显提高了准确性,因为直接注入了先验知识以提高其鲁棒性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号