首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing;ICASSP >Scalable stacking and learning for building deep architectures
【24h】

Scalable stacking and learning for building deep architectures

机译:可扩展的堆栈和学习以构建深度架构

获取原文

摘要

Deep Neural Networks (DNNs) have shown remarkable success in pattern recognition tasks. However, parallelizing DNN training across computers has been difficult. We present the Deep Stacking Network (DSN), which overcomes the problem of parallelizing learning algorithms for deep architectures. The DSN provides a method of stacking simple processing modules in buiding deep architectures, with a convex learning problem in each module. Additional fine tuning further improves the DSN, while introducing minor non-convexity. Full learning in the DSN is batch-mode, making it amenable to parallel training over many machines and thus be scalable over the potentially huge size of the training data. Experimental results on both the MNIST (image) and TIMIT (speech) classification tasks demonstrate that the DSN learning algorithm developed in this work is not only parallelizable in implementation but it also attains higher classification accuracy than the DNN.
机译:深度神经网络(DNN)在模式识别任务中显示出了惊人的成功。但是,在计算机之间并行化DNN训练非常困难。我们提出了深度堆栈网络(DSN),该网络克服了针对深度架构并行化学习算法的问题。 DSN提供了一种在深度架构中堆叠简单处理模块的方法,每个模块中都有一个凸学习问题。附加的微调进一步改善了DSN,同时引入了较小的非凸性。 DSN中的完全学习是批处理模式,使其可以在许多机器上进行并行训练,因此可以在可能巨大的训练数据规模上进行扩展。在MNIST(图像)和TIMIT(语音)分类任务上的实验结果表明,这项工作中开发的DSN学习算法不仅在实现上可并行化,而且比DNN具有更高的分类精度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号