首页> 外文期刊>Language, cognition and neuroscience >Task-based evaluation of context-sensitive referring expressions in human-robot dialogue
【24h】

Task-based evaluation of context-sensitive referring expressions in human-robot dialogue

机译:基于任务的人机对话中上下文相关引用表达的评估

获取原文
获取原文并翻译 | 示例
           

摘要

The standard referring-expression generation task involves creating stand-alone descriptions intended solely to distinguish a target object from its context. However, when an artificial system refers to objects in the course of interactive, embodied dialogue with a human partner, this is a very different setting; the references found in situated dialogue are able to take into account the aspects of the physical, interactive and task-level context, and are therefore unlike those found in corpora of standalone references. Also, the dominant method of evaluating generated references involves measuring corpus similarity. In an interactive context, though, other extrinsic measures such as task success and user preference are more relevant - and numerous studies have repeatedly found little or no correlation between such extrinsic metrics and the predictions of commonly used corpus-similarity metrics. To explore these issues, we introduce a humanoid robot designed to cooperate with a human partner on a joint construction task. We then describe the context-sensitive reference-generation algorithm that was implemented for use on this robot, which was inspired by the referring phenomena found in the Joint Construction Task corpus of human-human joint construction dialogues. The context-sensitive algorithm was evaluated through two user studies comparing it to a baseline algorithm, using a combination of objective performance measures and subjective user satisfaction scores. In both studies, the objective task performance and dialogue quality were found to be the same for both versions of the system; however, in both cases, the context-sensitive system scored more highly on subjective measures of interaction quality.
机译:标准的引用表达生成任务涉及创建独立的描述,这些描述仅旨在将目标对象与其上下文区分开。但是,当人工系统在与人类伙伴进行互动,具体化对话的过程中引用对象时,情况就大不相同了;在对话中找到的参考文献能够考虑物理,交互和任务级别上下文的各个方面,因此与独立参考文献集中的参考文献不同。同样,评估生成的参考文献的主要方法涉及测量语料库相似度。但是,在交互式环境中,其他外部措施(例如任务成功和用户偏好)更加相关-大量研究反复发现,这些外部度量与常用语料相似度度量的预测之间几乎没有相关性。为了探讨这些问题,我们引入了一种拟人机器人,该机器人旨在与人类伙伴合作完成共同的建造任务。然后,我们描述了在此机器人上使用的上下文相关参考生成算法,该算法的灵感来自于人与人共同建设对话的共同建设任务语料库中的参照现象。通过两次用户研究对上下文敏感算法进行了评估,并结合了客观性能指标和主观用户满意度评分,将其与基线算法进行了比较。在两项研究中,发现系统的两个版本的客观任务绩效和对话质量均相同;但是,在这两种情况下,上下文相关系统在交互质量的主观度量上得分都更高。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号