首页> 外文期刊>Image and Vision Computing >Development and utilization of a disgusting image dataset to understand and predict visual disgust
【24h】

Development and utilization of a disgusting image dataset to understand and predict visual disgust

机译:开发和利用令人反感的图像数据集来理解和预测视觉反感

获取原文
获取原文并翻译 | 示例
           

摘要

When viewers search for images on the internet, they may unexpectedly encounter disgusting or explicit images. As such images may result in mental suffering or trauma, predicting whether images will induce disgust in order to avoid such issues is desirable. However, formal definitions or insights as to what constitutes disgust-inducing visual factors do not exist. Consequently, eliminating disgusting images from retrieval results is still a challenge. In this paper, we collect a large-scale disgust-inducing image dataset containing approximately 60,000 images, each labeled with disgust scores and divided into image categories. Subsequently, using our dataset, we explore various attributes of disgust-inducing images, such as score distributions, categories of disgusting images, and relationships with other visual attributes. Then, we develop a new Convolutional Neural Network (CNN), called DiNet, that uses more than two pre-trained convolutional layers to consider local to global features for image representation. Experimental results indicate that the developed CNN architecture outperforms both feature-based learning models and state-of-the-art deep learning models with an accuracy of 67.58%. Furthermore, disgust maps extracted using the developed model facilitate an understanding of the disgust-inducing regions of images. (C) 2018 Elsevier B.V. All rights reserved.
机译:观看者在Internet上搜索图像时,可能会意外地遇到令人恶心或露骨的图像。由于此类图像可能会导致精神痛苦或创伤,因此需要预测图像是否会引起厌恶,从而避免此类问题。然而,关于什么构成令人反感的视觉因素的正式定义或见解并不存在。因此,从检索结果中消除令人反感的图像仍然是一个挑战。在本文中,我们收集了一个大规模的令人反感的图像数据集,其中包含约60,000张图像,每张图像均标有令人反感的评分并分为图像类别。随后,使用我们的数据集,探索令人反感的图像的各种属性,例如得分分布,令人反感的图像的类别以及与其他视觉属性的关系。然后,我们开发了一个称为DiNet的新卷积神经网络(CNN),该网络使用了两个以上的预训练卷积层来考虑局部到全局特征以进行图像表示。实验结果表明,所开发的CNN架构的性能优于基于特征的学习模型和最先进的深度学习模型,其准确性为67.58%。此外,使用开发的模型提取的反感图有助于理解图像的反感区域。 (C)2018 Elsevier B.V.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号