...
首页> 外文期刊>Neurocomputing >Dropout regularization in hierarchical mixture of experts
【24h】

Dropout regularization in hierarchical mixture of experts

机译:专家的分层混合中的辍学正常化

获取原文
获取原文并翻译 | 示例
           

摘要

Dropout is a very effective method in preventing overfitting and has become the go-to regularizer for multi-layer neural networks in recent years. Hierarchical mixture of experts is a hierarchically gated model that defines a soft decision tree where leaves correspond to experts and decision nodes correspond to gating models that softly choose between its children, and as such, the model defines a soft hierarchi-cal partitioning of the input space. In this work, we propose a variant of dropout for hierarchical mixture of experts that is faithful to the tree hierarchy defined by the model, as opposed to having a flat, unitwise independent application of dropout as one has with multi-layer perceptrons. We show that on a synthetic regression data and on MNIST, CIFAR-10, and SSTB datasets, our proposed dropout mechanism prevents overfitting on trees with many levels improving generalization and providing smoother fits. (c) 2020 Elsevier B.V. All rights reserved.
机译:辍学是一种非常有效的方法,防止过度装备,并且已成为近年来多层神经网络的转向规律器。专家的分层混合是一个分层门控模型,它定义了一个软决策树,其中叶对应于专家和决策节点对应于在其子节点之间轻声选择的门控模型,因此,该模型定义了输入的软层次结构分区空间。在这项工作中,我们提出了一种辍学的变体,用于忠于模型定义的树层次,而不是具有平坦,单独独立于辍学的辍学,因为具有多层的影响。我们展示在合成回归数据和MNIST,CIFAR-10和SSTB数据集上,我们提出的辍学机制可防止树木上的过度装备,其中许多级别改善泛化并提供更平滑的拟合。 (c)2020 Elsevier B.v.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号