首页> 外文会议>Annual meeting of the special interest group on discourse and dialogue >Multimodal Hierarchical Reinforcement Learning Policy for Task-Oriented Visual Dialog
【24h】

Multimodal Hierarchical Reinforcement Learning Policy for Task-Oriented Visual Dialog

机译:面向任务的可视对话框的多模式分层强化学习策略

获取原文

摘要

Creating an intelligent conversational system that understands vision and language is one of the ultimate goals in Artificial Intelligence (AI) (Winograd, 1972). Extensive research has focused on vision-to-language generation, however, limited research has touched on combining these two modalities in a goal-driven dialog context. We propose a multimodal hierarchical reinforcement learning framework that dynamically integrates vision and language for task-oriented visual dialog. The framework jointly learns the multimodal dialog state representation and the hierarchical dialog policy to improve both dialog task success and efficiency. We also propose a new technique, state adaptation, to integrate context awareness in the dialog state representation. We evaluate the proposed framework and the state adaptation technique in an image guessing game and achieve promising results.
机译:创建能够理解视觉和语言的智能对话系统是人工智能(AI)的最终目标之一(Winograd,1972)。广泛的研究集中在视觉到语言的生成上,但是,有限的研究涉及在目标驱动的对话上下文中结合这两种方式。我们提出了一种多模式的分层强化学习框架,该框架可动态集成视觉和语言以实现面向任务的视觉对话。该框架共同学习多模式对话框状态表示和分层对话框策略,以提高对话框任务的成功率和效率。我们还提出了一种新的技术,即状态适应,以将上下文意识集成到对话框状态表示中。我们在图像猜测游戏中评估了提出的框架和状态适应技术,并取得了可喜的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号