首页> 外文会议>Workshop on Human Evaluation of NLP Systems >Estimating Subjective Crowd-Evaluations as an Additional Objective to Improve Natural Language Generation
【24h】

Estimating Subjective Crowd-Evaluations as an Additional Objective to Improve Natural Language Generation

机译:估算主观人群评估作为提高自然语言生成的额外目标

获取原文

摘要

Human ratings are one of the most prevalent methods to evaluate the performance of natural language processing algorithms. Similarly, it is common to measure the quality of sentences generated by a natural language generation model using human raters. In this paper, we argue for exploring the use of subjective evaluations within the process of training language generation models in a multi-task learning setting. As a case study, we use a crowd-authored dialogue corpus to fine-tune six different language generation models. Two of these models incorporate multi-task learning and use subjective ratings of lines as part of an explicit learning goal. A human evaluation of the generated dialogue lines reveals that utterances generated by the multi-tasking models were subjectively rated as the most typical, most moving the conversation forward, and least offensive. Based on these promising first results, we discuss future research directions for incorporating subjective human evaluations into language model training and to hence keep the human user in the loop during the development process.
机译:人类评级是评估自然语言处理算法性能的最普遍的方法之一。类似地,通常使用人类评估者测量由自然语言生成模型产生的句子的质量。在本文中,我们争辩探讨在多任务学习设置中培训语言生成模型过程中使用主观评估的使用。作为一个案例研究,我们使用人群撰写的对话语料库来微调六种不同的语言生成模型。这些模型中的两个包含多任务学习,并使用线的主观评级作为明确学习目标的一部分。对生成的对话线的人类评估表明,由多任务模型产生的话语是最受主观的评级,作为最典型的,最大的对话前进,最不令人反感。基于这些有前途的第一个结果,我们讨论了将主观人类评估纳入语言模型培训的未来研究方向,从而在开发过程中将人类用户保持在循环中。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号