首页> 外文会议>International Conference on Quality of Multimedia Experience >A Crowdsourcing Approach to Evaluate the Quality of Query-based Extractive Text Summaries
【24h】

A Crowdsourcing Approach to Evaluate the Quality of Query-based Extractive Text Summaries

机译:一种评价基于查询的提取文本摘要质量的众群方法

获取原文

摘要

High cost and time consumption are concurrent barriers for research and application of automated summarization. In order to explore options to overcome this barrier, we analyze the feasibility and appropriateness of micro-task crowdsourcing for evaluation of different summary quality characteristics and report an ongoing work on the crowdsourced evaluation of query-based extractive text summaries. To do so, we assess and evaluate a number of linguistic quality factors such as grammaticality, non-redundancy, referential clarity, focus and structure & coherence. Our first results imply that referential clarity, focus and structure & coherence are the main factors effecting the perceived summary quality by crowdworkers. Further, we compare these results using an initial set of expert annotations that is currently being collected, as well as an initial set of automatic quality score ROUGE for summary evaluation. Preliminary results show that ROUGE does not correlate with linguistic quality factors, regardless if assessed by crowd or experts. Further, crowd and expert ratings show highest degree of correlation when assessing low quality summaries. Assessments increasingly divert when attributing high quality judgments.
机译:高成本和时间消耗是用于自动摘要的研究和应用的并发障碍。为了探索克服这一屏障的选择,我们分析了微任众众覆盖的可行性和适当性,以评估不同的总结质量特征,并报告了对基于查询的提取文本摘要的众群评估的持续工作。为此,我们评估并评估许多语言质量因素,如语法,非冗余,指称清晰度,焦点和结构和一致性。我们的第一款结果意味着指称清晰度,焦点和结构和连贯性是群体影响人群的主要因素。此外,我们使用当前正在收集的初始专家注释以及概要的摘要评估的初始自动质量分数胭脂集进行比较这些结果。初步结果表明,无论人群或专家评估,胭脂与语言质量因素无关。此外,当评估低质量摘要时,人群和专家评级显示出最高的相关程度。评估在归因于归因于高质量的判断时日益转移。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号