首页> 外文会议>Conference on Optics and photonics for counterterrorism, crime fighting and defence XII >Object recognition using deep convolutional neural networks with complete transfer and partial frozen layers
【24h】

Object recognition using deep convolutional neural networks with complete transfer and partial frozen layers

机译:对象识别使用完全转移和部分冻结的深度卷积神经网络

获取原文

摘要

Object recognition is important to understand the content of video and allow flexible querying in a large number of cameras, especially for security applications. Recent benchmarks show that deep convolutional neural networks are excellent approaches for object recognition. This paper describes an approach of domain transfer, where features learned from a large annotated dataset are transferred to a target domain where less annotated examples are available as is typical for the security and defense domain. Many of these networks trained on natural images appear to learn features similar to Gabor filters and color blobs in the first layer. These first-layer features appear to be generic for many datasets and tasks while the last layer is specific. In this paper, we study the effect of copying all layers and fine-tuning a variable number. We performed an experiment with a Caffe-based network on 1000 ImageNet classes that are randomly divided in two equal subgroups for the transfer from one to the other. We copy all layers and vary the number of layers that is fine-tuned and the size of the target dataset. We performed additional experiments with the Keras platform on CIFAR-10 dataset to validate general applicability. We show with both platforms and both datasets that the accuracy on the target dataset improves when more target data is used. When the target dataset is large, it is beneficial to freeze only a few layers. For a large target dataset, the network without transfer learning performs better than the transfer network, especially if many layers are frozen. When the target dataset is small, it is beneficial to transfer (and freeze) many layers. For a small target dataset, the transfer network boosts generalization and it performs much better than the network without transfer learning. Learning time can be reduced by freezing many layers in a network.
机译:对象识别对于了解视频的内容非常重要,并允许灵活地查询大量相机,尤其是对于安全应用程序。最近的基准表明,深度卷积神经网络是对象识别的优异方法。本文介绍了域传输的方法,其中从大型注释数据集中学习的特征被传送到目标域,其中可用于安全性和防御域的典型值较少的注释示例。许多在自然图像上培训的这些网络似乎学习了类似于Gabor滤波器和第一层颜色斑点的功能。这些第一层特征对于许多数据集和任务似乎是通用的,而最后一层是特定的。在本文中,我们研究了复制所有层和微调变量数的效果。我们在1000个ImageNet类上进行了一个基于Caffe的网络进行了实验,这些网络在两个相等的子组中随机划分为从一个到另一个的转移。我们复制所有图层并改变微调的层数和目标数据集的大小。我们对CiFar-10数据集的Keras平台进行了额外的实验,以验证一般适用性。我们以平台和两个数据集显示在使用更多目标数据时,目标数据集的准确性提高。当目标数据集很大时,只冻结几层是有益的。对于大型目标数据集,无需传输学习的网络比传输网络更好地执行,特别是如果多层被冻结。当目标数据集很小时,它有利于转移(和冻结)许多层。对于小型目标数据集,传输网络提升泛化,而且在没有传输学习的情况下执行比网络更好。通过在网络中冻结许多层来减少学习时间。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号