首页> 外文会议>International Conference on Quality of Multimedia Experience >A Crowdsourcing Approach to Evaluate the Quality of Query-based Extractive Text Summaries
【24h】

A Crowdsourcing Approach to Evaluate the Quality of Query-based Extractive Text Summaries

机译:一种评估基于查询的提取文本摘要质量的众包方法

获取原文

摘要

High cost and time consumption are concurrent barriers for research and application of automated summarization. In order to explore options to overcome this barrier, we analyze the feasibility and appropriateness of micro-task crowdsourcing for evaluation of different summary quality characteristics and report an ongoing work on the crowdsourced evaluation of query-based extractive text summaries. To do so, we assess and evaluate a number of linguistic quality factors such as grammaticality, non-redundancy, referential clarity, focus and structure & coherence. Our first results imply that referential clarity, focus and structure & coherence are the main factors effecting the perceived summary quality by crowdworkers. Further, we compare these results using an initial set of expert annotations that is currently being collected, as well as an initial set of automatic quality score ROUGE for summary evaluation. Preliminary results show that ROUGE does not correlate with linguistic quality factors, regardless if assessed by crowd or experts. Further, crowd and expert ratings show highest degree of correlation when assessing low quality summaries. Assessments increasingly divert when attributing high quality judgments.
机译:高成本和时间消耗是自动汇总研究和应用的并发障碍。为了探索克服此障碍的选项,我们分析了微任务众包用于评估不同的摘要质量特征的可行性和适当性,并报告了基于查询的提取文本摘要的众包评估的正在进行的工作。为此,我们评估和评估许多语言质量因素,例如语法,非冗余,参照清晰度,重点和结构与连贯性。我们的第一个结果表明,参照的清晰性,重点,结构和连贯性是影响人群工作者感知的摘要质量的主要因素。此外,我们使用当前正在收集的一组初始专家注释以及一组用于摘要评估的自动质量得分ROUGE初始组来比较这些结果。初步结果表明,无论是通过人群还是专家进行评估,ROUGE与语言质量因素均不相关。此外,在评估低质量摘要时,人群和专家评分显示出最高的相关度。归因于高质量的判断时,评估越来越转移。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号