首页> 外文会议>IEEE Automatic Speech Recognition and Understanding Workshop >Dialogue Environments are Different from Games: Investigating Variants of Deep Q-Networks for Dialogue Policy
【24h】

Dialogue Environments are Different from Games: Investigating Variants of Deep Q-Networks for Dialogue Policy

机译:对话环境与游戏不同:对对话政策进行深度Q网络的调查变体

获取原文

摘要

The dialogue manager is an important component in a task-oriented dialogue system, which focuses on deciding dialogue policy given the dialogue state in order to fulfill the user goal. Learning dialogue policy is usually framed as a reinforcement learning (RL) problem, where the objective is to maximize the reward indicating whether the conversation is successful and how efficient it is. However, even there are many variants of deep Q-networks (DQN) achieving better performance on game playing scenarios, no prior work analyzed the performance of dialogue policy learning using these improved versions. Considering that dialogue interactions differ a lot from game playing, this paper investigates variants of DQN models together with different exploration strategies in a benchmark experimental setup, and then we examine which RL methods are more suitable for task-completion dialogue policy learning11The code is available at https://github.com/MiuLab/Dialogue-DQN-Variants.
机译:对话经理是一项面向任务导向的对话系统的重要组成部分,专注于为履行对话状态而定的对话政策,以满足用户目标。学习对话政策通常被诬陷为加强学习(RL)问题,其中目标是最大限度地提高奖励,表明谈话是否成功,它有多高效。然而,即使有许多深度Q-Networks(DQN)的变种,在游戏场景中实现了更好的性能,也没有先前的工作分析了使用这些改进版本的对话策略学习的性能。考虑到对话互动不同,从游戏播放中有很多差异,本文将DQN模型的变体与基准实验设置中的不同探索策略一起调查,然后我们检查哪些RL方法更适合任务完成对话政策学习 1 1 代码可在https://github.com/miulab/dialogue-dqn-variants获得。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号