首页> 外文期刊>Big Data >A Literature Survey and Experimental Evaluation of the State-of-the-Art in Uplift Modeling: A Stepping Stone Toward the Development of Prescriptive Analytics
【24h】

A Literature Survey and Experimental Evaluation of the State-of-the-Art in Uplift Modeling: A Stepping Stone Toward the Development of Prescriptive Analytics

机译:文献综述和对提升模型研究的最新实验评估:规范分析发展的垫脚石

获取原文
获取原文并翻译 | 示例
           

摘要

Prescriptive analytics extends on predictive analytics by allowing to estimate an outcome in function of control variables, allowing as such to establish the required level of control variables for realizing a desired outcome. Uplift modeling is at the heart of prescriptive analytics and aims at estimating the net difference in an outcome resulting from a specific action or treatment that is applied. In this article, a structured and detailed literature survey on uplift modeling is provided by identifying and contrasting various groups of approaches. In addition, evaluation metrics for assessing the performance of uplift models are reviewed. An experimental evaluation on four real-world data sets provides further insight into their use. Uplift random forests are found to be consistently among the best performing techniques in terms of the Qini and Gini measures, although considerable variability in performance across the various data sets of the experiments is observed. In addition, uplift models are frequently observed to be unstable and display a strong variability in terms of performance across different folds in the cross-validation experimental setup. This potentially threatens their actual use for business applications. Moreover, it is found that the available evaluation metrics do not provide an intuitively understandable indication of the actual use and performance of a model. Specifically, existing evaluation metrics do not facilitate a comparison of uplift models and predictive models and evaluate performance either at an arbitrary cutoff or over the full spectrum of potential cutoffs. In conclusion, we highlight the instability of uplift models and the need for an application-oriented approach to assess uplift models as prime topics for further research.
机译:规范分析通过允许根据控制变量的功能来估计结果,从而允许建立所需的控制变量级别以实现期望的结果,从而扩展了预测分析。提升建模是说明性分析的核心,旨在估算由所应用的特定操作或治疗导致的结果的净差异。在本文中,通过识别和对比各种方法组,提供了关于隆升模型的结构化且详细的文献调查。此外,审查了评估提升模型性能的评估指标。对四个真实数据集的实验评估提供了对其使用的进一步了解。尽管在实验的各种数据集之间观察到性能差异很大,但就奇尼和基尼度量而言,隆起随机森林一直被认为是性能最好的技术之一。此外,在交叉验证实验设置中,经常会观察到隆起模型不稳定并且在跨不同折痕的性能方面表现出很大的可变性。这可能会威胁到它们在业务应用程序中的实际使用。而且,发现可用的评估度量不能提供模型的实际使用和性能的直观理解。具体而言,现有的评估指标不利于对提升模型和预测模型进行比较,也无法在任意临界值或整个潜在临界值范围内评估性能。总之,我们强调了隆升模型的不稳定性以及需要一种面向应用的方法来评估隆升模型,并将其作为进一步研究的主要主题。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号