【24h】

Affective Visual Question Answering Network

机译:情感视觉问答网络

获取原文
获取原文并翻译 | 示例

摘要

Visual Question Answering (VQA) has recently attracted considerable attention from researchers in the trending field of deep learning. The need to improve VQA models by focusing on local regions of images, has resulted in the development of various attention models. This paper proposes the Affective Visual Question Answering Network (AVQAN), an attention model that combines the locality of the image features, the question and the mood detected from the specific region of the image to produce an affective answer using a preprocessed image dataset. The experimental results depict that AVQAN enriches the analysis and understanding of images by adding affective information to the answer, while still managing to maintain the accuracy levels within the range of recent ordinary VQA baseline models. The proposed model significantly contributes towards the development of rapidly improving emotion-aware machines that are becoming increasingly vital in everyday life.
机译:视觉问答(VQA)最近已经引起了深度学习趋势领域研究人员的极大关注。通过关注图像的局部区域来改进VQA模型的需求导致了各种注意力模型的开发。本文提出了一种情感视觉问题解答网络(AVQAN),该注意力模型结合了图像特征的局部性,问题和从图像特定区域检测到的情绪,从而使用预处理的图像数据集产生了情感答案。实验结果表明,AVQAN通过向答案中添加情感信息来丰富图像的分析和理解,同时仍设法将准确度保持在最近的普通VQA基线模型范围内。所提出的模型极大地促进了迅速改善的情绪感知机器的发展,这些机器在日常生活中变得越来越重要。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号