首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Scalable stacking and learning for building deep architectures
【24h】

Scalable stacking and learning for building deep architectures

机译:用于构建深层架构的可扩展堆叠和学习

获取原文

摘要

Deep Neural Networks (DNNs) have shown remarkable success in pattern recognition tasks. However, parallelizing DNN training across computers has been difficult. We present the Deep Stacking Network (DSN), which overcomes the problem of parallelizing learning algorithms for deep architectures. The DSN provides a method of stacking simple processing modules in buiding deep architectures, with a convex learning problem in each module. Additional fine tuning further improves the DSN, while introducing minor non-convexity. Full learning in the DSN is batch-mode, making it amenable to parallel training over many machines and thus be scalable over the potentially huge size of the training data. Experimental results on both the MNIST (image) and TIMIT (speech) classification tasks demonstrate that the DSN learning algorithm developed in this work is not only parallelizable in implementation but it also attains higher classification accuracy than the DNN.
机译:深度神经网络(DNN)在模式识别任务中显示出显着的成功。但是,跨电脑的DNN培训并行化已经很困难。我们介绍了深度堆叠网络(DSN),克服了对深层架构并行化学习算法的问题。 DSN提供了一种堆叠简单处理模块的方法,在构建深度架构中,每个模块中的凸起学习问题。额外的微调进一步改善了DSN,同时引入了轻微的非凸性。 DSN中的全面学习是批量模式,使其适用于许多机器上的并行训练,因此可以通过培训数据的潜在大小进行可扩展。 MNIST(图像)和时间(语音)分类任务的实验结果表明,在该工作中开发的DSN学习算法不仅在实施中并行,而且还比DNN达到更高的分类精度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号