...
首页> 外文期刊>Computational intelligence and neuroscience >An Improved Convolutional Neural Network Algorithm and Its Application in Multilabel Image Labeling
【24h】

An Improved Convolutional Neural Network Algorithm and Its Application in Multilabel Image Labeling

机译:一种改进的卷积神经网络算法及其在多标签图像标记中的应用

获取原文
   

获取外文期刊封面封底 >>

       

摘要

In today’s society, image resources are everywhere, and the number of available images can be overwhelming. Determining how to rapidly and effectively query, retrieve, and organize image information has become a popular research topic, and automatic image annotation is the key to text-based image retrieval. If the semantic images with annotations are not balanced among the training samples, the low-frequency labeling accuracy can be poor. In this study, a dual-channel convolution neural network (DCCNN) was designed to improve the accuracy of automatic labeling. The model integrates two convolutional neural network (CNN) channels with different structures. One channel is used for training based on the low-frequency samples and increases the proportion of low-frequency samples in the model, and the other is used for training based on all training sets. In the labeling process, the outputs of the two channels are fused to obtain a labeling decision. We verified the proposed model on the Caltech-256, Pascal VOC 2007, and Pascal VOC 2012 standard datasets. On the Pascal VOC 2012 dataset, the proposed DCCNN model achieves an overall labeling accuracy of up to 93.4% after 100 training iterations: 8.9% higher than the CNN and 15% higher than the traditional method. A similar accuracy can be achieved by the CNN only after 2,500 training iterations. On the 50,000-image dataset from Caltech-256 and Pascal VOC 2012, the performance of the DCCNN is relatively stable; it achieves an average labeling accuracy above 93%. In contrast, the CNN reaches an accuracy of only 91% even after extended training. Furthermore, the proposed DCCNN achieves a labeling accuracy for low-frequency words approximately 10% higher than that of the CNN, which further verifies the reliability of the proposed model in this study.
机译:在当今的社会中,图像资源无处不在,可用图像的数量可能是压倒性的。确定如何快速有效地查询,检索和组织图像信息已成为流行的研究主题,自动图像注释是基于文本的图像检索的关键。如果在训练样本中没有注释的语义图像不平衡,则低频标记精度可能差。在本研究中,设计了一种双通道卷积神经网络(DCCNN),旨在提高自动标签的准确性。该模型与不同的结构集成了两个卷积神经网络(CNN)通道。一个通道用于基于低频样本的培训,并增加模型中的低频样本的比例,另一个通道增加了基于所有训练集的培训。在标签过程中,两个通道的输出被融合以获得标记决定。我们验证了CALTECH-256,Pascal VOC 2007和Pascal VOC 2012标准数据集的拟议模型。在Pascal VOC 2012 DataSet上,拟议的DCCNN模型在100次训练迭代后达到高达93.4%的总标记精度:8.9%高于CNN,比传统方法高15%。 CNN仅在2,500次训练迭代之后可以通过CNN实现类似的准确度。在Caltech-256和Pascal VOC 2012的50,000图像数据集上,DCCNN的性能相对稳定;它实现了高于93%以上的平均标记精度。相比之下,即使在扩展训练之后,CNN即使在延长训练后也达到了91%的准确性。此外,所提出的DCCNN用于低频率字的标签精度比CNN高约10%,这进一步验证了本研究中所提出的模型的可靠性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号