首页> 外文期刊>Machine Learning >CaDET: interpretable parametric conditional density estimation with decision trees and forests
【24h】

CaDET: interpretable parametric conditional density estimation with decision trees and forests

机译:CaDET:具有决策树和森林的可解释参数条件密度估计

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

We introduce CaDET, an algorithm for parametric Conditional Density Estimation (CDE) based on decision trees and random forests. CaDET uses the empirical cross entropy impurity criterion for tree growth, which incentivizes splits that improve predictive accuracy more than the regression criteria or estimated mean-integrated-square-error used in previous works. CaDET also admits more efficient training and query procedures than existing tree-based CDE approaches, and stores only a bounded amount of information at each tree leaf, by using sufficient statistics for all computations. Previous tree-based CDE techniques produce complicated uninterpretable distribution objects, whereas CaDET may be instantiated with easily interpretable distribution families, making every part of the model easy to understand. Our experimental evaluation on real datasets shows that CaDET usually learns more accurate, smaller, and more interpretable models, and is less prone to overfitting than existing tree-based CDE approaches.
机译:我们介绍CaDET,一种基于决策树和随机森林的参数条件密度估计(CDE)算法。 CaDET使用经验性的交叉熵杂质准则进行树木生长,该准则可激励劈裂,比以前的工作中使用的回归准则或估计的均方误差,其预测准确性更高。与现有的基于树的CDE方法相比,CaDET还接受了更有效的训练和查询过程,并且通过对所有计算使用足够的统计信息,CaDET仅在每个树叶中存储有限量的信息。以前的基于树的CDE技术会生成复杂的无法解释的分发对象,而CaDET可以使用易于解释的分发族实例化,从而使模型的每个部分都易于理解。我们对真实数据集的实验评估表明,与现有的基于树的CDE方法相比,CaDET通常可以学习更准确,更小,更易解释的模型,并且不太容易过度拟合。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号