【24h】

On Deriving the Second-Stage Training Set for Trainable Combiners

机译:关于推导可训练合成器第二阶段训练集的信息

获取原文
获取原文并翻译 | 示例

摘要

Unlike fixed combining rules, the trainable combiner is applicable to ensembles of diverse base classifier architectures with incomparable outputs. The trainable combiner, however, requires the additional step of deriving a second-stage training dataset from the base classifier outputs. Although several strategies have been devised, it is thus far unclear which is superior for a given situation. In this paper we investigate three principal training techniques, namely the re-use of the training dataset for both stages, an independent validation set, and the stacked generalization. On experiments with several datasets we have observed that the stacked generalization outperforms the other techniques in most situations, with the exception of very small sample sizes, in which the re-using strategy behaves better. We illustrate that the stacked generalization introduces additional noise to the second-stage training dataset, and should therefore be bundled with simple combiners that are insensitive to the noise. We propose an extension of the stacked generalization approach which significantly improves the combiner robustness.
机译:与固定组合规则不同,可训练组合器适用于具有无可比拟的输出的各种基础分类器体系结构的集合。但是,可训练的组合器需要额外的步骤,即从基本分类器的输出中得出第二阶段的训练数据集。尽管已经设计了几种策略,但是到目前为止,对于给定情况哪种策略更好尚不清楚。在本文中,我们研究了三种主要的训练技术,即在两个阶段都重复使用训练数据集,独立的验证集和堆叠概括。在使用多个数据集进行的实验中,我们观察到在大多数情况下,堆栈式概括优于其他技术,但样本量非常小(在这种情况下,重用策略的表现更好)。我们说明,堆叠概括将额外的噪声引入第二阶段的训练数据集,因此应与对噪声不敏感的简单组合器捆绑在一起。我们提出了一种堆叠泛化方法的扩展,可以显着提高组合器的鲁棒性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号