首页> 外文会议>International Conference on Computer Vision >SegEQA: Video Segmentation Based Visual Attention for Embodied Question Answering
【24h】

SegEQA: Video Segmentation Based Visual Attention for Embodied Question Answering

机译:Segeqa:基于视频分割的视觉关注,体现了问题应答

获取原文
获取外文期刊封面目录资料

摘要

Embodied Question Answering (EQA) is a newly defined research area where an agent is required to answer the user's questions by exploring the real world environment. It has attracted increasing research interests due to its broad applications in automatic driving system, in-home robots, and personal assistants. Most of the existing methods perform poorly in terms of answering and navigation accuracy due to the absence of local details and vulnerability to the ambiguity caused by complicated vision conditions. To tackle these problems, we propose a segmentation based visual attention mechanism for Embodied Question Answering. Firstly, We extract the local semantic features by introducing a novel high-speed video segmentation framework. Then by the guide of extracted semantic features, a bottom-up visual attention mechanism is proposed for the Visual Question Answering (VQA) sub-task. Further, a feature fusion strategy is proposed to guide the training of the navigator without much additional computational cost. The ablation experiments show that our method boosts the performance of VQA module by 4.2% (68.99% vs 64.73%) and leads to 3.6% (48.59% vs 44.98%) overall improvement in EQA accuracy.
机译:体现的问题回答(EQA)是一个新定义的研究领域,需要通过探索现实世界环境来回答用户的问题。由于其在自动驾驶系统,家庭机器人和个人助理中的广泛应用,它引起了越来越多的研究兴趣。由于缺乏由复杂的视觉条件引起的歧义的局部细节和脆弱,大多数现有方法在应答和导航准确性方面表现不佳。为了解决这些问题,我们提出了一种基于分段的视觉注意机制,以实现体现的问题。首先,我们通过引入新颖的高速视频分段框架来提取局部语义特征。然后通过提取的语义特征指南,提出了一个自下而上的视觉注意机制,用于视觉问题应答(VQA)子任务。此外,提出了一种特征融合策略,以指导导航器的训练,而无需额外的计算成本。消融实验表明,我们的方法将VQA模块的性能提高了4.2%(68.99%与64.73%),导致欧足联准确性的总体改善3.6%(48.59%vs44.98%)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号