首页> 外文期刊>Information retrieval >Pooling-based continuous evaluation of information retrieval systems
【24h】

Pooling-based continuous evaluation of information retrieval systems

机译:基于池的信息检索系统连续评估

获取原文
获取原文并翻译 | 示例
           

摘要

The dominant approach to evaluate the effectiveness of information retrieval (IR) systems is by means of reusable test collections built following the Cranfield paradigm. In this paper, we propose a new IR evaluation methodology based on pooled test-collections and on the continuous use of either crowdsourcing or professional editors to obtain relevance judgements. Instead of building a static collection for a finite set of systems known a priori, we propose an IR evaluation paradigm where retrieval approaches are evaluated iteratively on the same collection. Each new retrieval technique takes care of obtaining its missing relevance judgements and hence contributes to augmenting the overall set of relevance judgements of the collection. We also propose two metrics: Fairness Score, and opportunistic number of relevant documents, which we then use to define new pooling strategies. The goal of this work is to study the behavior of standard IR metrics, IR system ranking, and of several pooling techniques in a continuous evaluation context by comparing continuous and non-continuous evaluation results on classic test collections. We both use standard and crowdsourced relevance judgements, and we actually run a continuous evaluation campaign over several existing IR systems.
机译:评估信息检索(IR)系统有效性的主要方法是通过遵循Cranfield范式建立的可重用测试集合。在本文中,我们提出了一种新的IR评估方法,该方法基于汇总的测试集合并持续使用众包或专业编辑来获得相关性判断。代替为已知先验的有限系统建立静态集合,我们提出一种IR评估范式,其中对同一集合迭代地评估检索方法。每种新的检索技术都需要获得其缺失的相关性判断,因此有助于增加集合的相关性判断的整体集合。我们还提出了两个指标:公平性得分和相关文档的机会数量,然后我们将其用于定义新的合并策略。这项工作的目的是通过比较经典测试集合上的连续和非连续评估结果,研究标准IR指标,IR系统排名和几种合并技术在连续评估环境中的行为。我们都使用标准和众包相关性判断,实际上我们对几个现有的IR系统进行了持续评估。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号