首页> 外文会议>Conference on Artificial Intelligence in Medicine >Local vs. Global Interpretability of Machine Learning Models in Type 2 Diabetes Mellitus Screening
【24h】

Local vs. Global Interpretability of Machine Learning Models in Type 2 Diabetes Mellitus Screening

机译:2型糖尿病筛选中机器学习模型的本地与全球可解释性

获取原文

摘要

Machine learning based predictive models have been used in different areas of everyday life for decades. However, with the recent availability of big data, new ways emerge on how to interpret the decisions of machine learning models. In addition to global interpretation focusing on the general prediction model decisions, this paper emphasizes the importance of local interpretation of predictions. Local interpretation focuses on specifics of each individual and provides explanations that can lead to a better understanding of the feature contribution in smaller groups of individuals that are often overlooked by the global interpretation techniques. In this paper, three machine learning based prediction models were compared: Gradient Boosting Machine (GBM), Random Forest (RF) and Generalized linear model with regularization (GLM). No significant differences in prediction performance, measured by mean average error, were detected: GLM: 0.573 (0.569 - 0.577); GBM: 0.579 (0.575 - 0.583); RF: 0.579 (0.575 - 0.583). Similar to other studies that used prediction models for screening in type 2 diabetes mellitus, we found a strong contribution of features like age, gender and BMI on the global interpretation level. On the other hand, local interpretation technique discovered some features like depression, smoking status or physical activity that can be influential in specific groups of patients. This study outlines the prospects of using local interpretation techniques to improve the interpretability of prediction models in the era of personalized healthcare. At the same time, we try to warn the users and developers of prediction models that prediction performance should not be the only criteria for model selection.
机译:基于机器学习的预测模型已经在日常生活的不同领域使用了几十年。但是,随着最近的大数据的可用性,新方式出现了如何解释机器学习模型的决策。除了全球解释专注于一般预测模型决策外,本文强调了当地解释预测的重要性。本地解释侧重于每个人的细节,并提供解释,这可以导致更好地了解较小的全球解释技术常被忽视的较小群体的特征贡献。在本文中,比较了三种基于机器学习的预测模型:具有正则化(GLM)的梯度升压机(GBM),随机林(RF)和广义线性模型。检测到通过平均误差测量的预测性能的显着差异:GLM:0.573(0.569 - 0.577); GBM:0.579(0.575 - 0.583); RF:0.579(0.575 - 0.583)。类似于其他研究,其中使用用于筛选2型糖尿病的预测模型,我们发现了在全球解释水平上的年龄,性别和BMI等特征的强烈贡献。另一方面,局部解释技术发现了一些抑郁症,吸烟状态或身体活动等特征,这些功能可能在特定患者组中有影响力。本研究概述了利用本地解释技术的前景,以改善个性化医疗保健时代预测模型的可解释性。与此同时,我们试图警告用户和开发人员的预测模型,即预测性能不应该是模型选择的唯一标准。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号