首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing;ICASSP >Off-policy learning in large-scale POMDP-based dialogue systems
【24h】

Off-policy learning in large-scale POMDP-based dialogue systems

机译:基于POMDP的大型对话系统中的非政策学习

获取原文

摘要

Reinforcement learning (RL) is now part of the state of the art in the domain of spoken dialogue systems (SDS) optimisation. Most performant RL methods, such as those based on Gaussian Processes, require to test small changes in the policy to assess them as improvements or degradations. This process is called on policy learning. Nevertheless, it can result in system behaviours that are not acceptable by users. Learning algorithms should ideally infer an optimal strategy by observing interactions generated by a non-optimal but acceptable strategy, that is learning off-policy. Such methods usually fail to scale up and are thus not suited for real-world systems. In this contribution, a sample-efficient, online and off-policy RL algorithm is proposed to learn an optimal policy. This algorithm is combined to a compact non-linear value function representation (namely a multi-layers perceptron) enabling to handle large scale systems.
机译:强化学习(RL)现在是口语对话系统(SDS)优化领域中最新技术的一部分。大多数高性能的RL方法(例如基于高斯过程的方法)都需要测试策略中的细微变化,以将其评估为改进或降级。此过程称为策略学习。但是,它可能导致用户无法接受的系统行为。理想情况下,学习算法应该通过观​​察非最佳但可接受的策略(即学习非策略)所产生的交互作用来推断最佳策略。这种方法通常无法按比例放大,因此不适用于实际系统。在此贡献中,提出了一种样本有效的在线和非策略RL算法,以学习最佳策略。该算法被组合为紧凑的非线性值函数表示形式(即多层感知器),能够处理大规模系统。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号