首页> 外文会议>IEEE Workshop on Automatic Speech Recognition and Understanding >A METHOD FOR EVALUATING AND COMPARING USER SIMULATIONS: THE CRAMER-VON MESES DIVERGENCE
【24h】

A METHOD FOR EVALUATING AND COMPARING USER SIMULATIONS: THE CRAMER-VON MESES DIVERGENCE

机译:一种评估和比较用户模拟的方法:克拉默 - 冯思斯分歧

获取原文

摘要

Although user simulations are increasingly employed in the development and assessment of spoken dialog systems, there is no accepted method for evaluating user simulations. In this paper, we propose a novel quality measure for user simulations. We view a user simulation as a predictor of the performance of a dialog system, where per-dialog performance is measured with a domain-specific scoring function. The quality of the user simulation is measured as the divergence between the distribution of scores in real dialogs and simulated dialogs, and we argue that the Cramer-von Mises divergence is well-suited to this task. The technique is demonstrated on a corpus of real calls, and we present a table of critical values for practitioners to interpret the statistical significance of comparisons between user simulations.
机译:虽然在口头对话系统的开发和评估方面越来越多地使用用户仿真,但没有接受的方法来评估用户仿真。在本文中,我们提出了一种新的用户模拟质量措施。我们将用户仿真视为对话系统性能的预测,其中每个对话表现使用域特定的评分函数测量。用户仿真的质量被测量为实际对话框中分数分布与模拟对话框之间的分配,我们认为克莱默 - von ices发散非常适合这项任务。该技术在真正的呼叫语料库上进行了演示,我们为从业者提供了一个临界值表,以解释用户仿真之间的比较的统计显着性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号