首页> 外文期刊>Information retrieval >Implementing crowdsourcing-based relevance experimentation: an industrial perspective
【24h】

Implementing crowdsourcing-based relevance experimentation: an industrial perspective

机译:实施基于众包的相关性实验:行业角度

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

Crowdsourcing has emerged as a viable platform for conducting different types of relevance evaluation. The main reason behind this trend is that it makes possible to conduct experiments extremely fast, with good results at a low cost. However, like in any experiment, there are several implementation details that would make an experiment work or fail. To gather useful results, clear instructions, user interface guidelines, content quality, inter-rater agreement metrics, work quality, and worker feedback are important characteristics of a successful crowdsourcing experiment. Furthermore, designing and implementing experiments that require thousands or millions of labels is different than conducting small scale research investigations. In this paper we outline a framework for conducting continuous crowdsourcing experiments, emphasizing aspects that should be of importance for all sorts of tasks. We illustrate the value of characteristics that can impact the overall outcome using examples based on TREC, INEX, and Wikipedia data sets.
机译:众包已经成为进行不同类型相关性评估的可行平台。这种趋势背后的主要原因是,它可以以极快的速度进行实验,并以低成本获得良好的结果。但是,就像在任何实验中一样,有一些实现细节会使实验工作或失败。为了收集有用的结果,清晰的说明,用户界面指南,内容质量,评估者之间的协议指标,工作质量和工作人员反馈是成功的众包实验的重要特征。此外,设计和实施需要成千上万个标签的实验与进行小规模研究调查不同。在本文中,我们概述了进行连续众包实验的框架,强调了对于所有任务都应具有重要意义的方面。我们使用基于TREC,INEX和Wikipedia数据集的示例说明了可能影响总体结果的特征的价值。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号