首页> 外文OA文献 >Improved learning algorithms for restricted Boltzmann machines
【2h】

Improved learning algorithms for restricted Boltzmann machines

机译:受限玻尔兹曼机的改进学习算法

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。
获取外文期刊封面目录资料

摘要

A restricted Boltzmann machine (RBM) is often used as a building block for constructing deep neural networks and deep generative models which have gained popularity recently as one way to learn complex and large probabilistic models. In these deep models, it is generally known that the layer-wise pretraining of RBMs facilitates finding a more accurate model for the data. It is, hence, important to have an efficient learning method for RBM. The conventional learning is mostly performed using the stochastic gradients, often, with the approximate method such as contrastive divergence (CD) learning to overcome the computational difficulty. Unfortunately, training RBMs with this approach is known to be difficult, as learning easily diverges after initial convergence. This difficulty has been reported recently by many researchers. This thesis contributes important improvements that address the difficulty of training RBMs. Based on an advanced Markov-Chain Monte-Carlo sampling method called parallel tempering (PT), the thesis proposes a PT learning which can replace CD learning. In terms of both the learning performance and the computational overhead, PT learning is shown to be superior to CD learning through various experiments. The thesis also tackles the problem of choosing the right learning parameter by proposing a new algorithm, the adaptive learning rate, which is able to automatically choose the right learning rate during learning. A closer observation into the update rules suggested that learning by the traditional update rules is easily distracted depending on the representation of data sets. Based on this observation, the thesis proposes a new set of gradient update rules that are more robust to the representation of training data sets and the learning parameters. Extensive experiments on various data sets confirmed that the proposed rules indeed improve learning significantly. Additionally, a Gaussian-Bernoulli RBM (GBRBM) which is a variant of an RBM that can learn continuous real-valued data sets is reviewed, and the proposed improvements are tested upon it. The experiments showed that the improvements could also be made for GBRBMs.
机译:受限的Boltzmann机器(RBM)通常用作构建深度神经网络和深度生成模型的构建模块,最近这些神经网络和深度生成模型已成为学习复杂和大型概率模型的一种方法。在这些深层模型中,众所周知,RBM的分层预训练有助于为数据找到更准确的模型。因此,重要的是有一种有效的RBM学习方法。传统的学习通常是使用随机梯度进行的,通常使用近似方法(例如,对比散度(CD)学习)来克服计算困难。不幸的是,众所周知,用这种方法来训练RBM是困难的,因为在初始收敛后学习容易分散。最近,许多研究人员已经报道了这种困难。本论文为解决RBM训练的困难做出了重要的贡献。基于一种称为并行回火(PT)的先进马尔可夫链蒙特卡洛采样方法,本文提出了一种可以代替CD学习的PT学习。在学习性能和计算开销方面,通过各种实验显示,PT学习优于CD学习。本文还提出了一种自适应学习率算法,该算法可以在学习过程中自动选择正确的学习率,从而解决了选择正确的学习参数的问题。对更新规则的仔细观察表明,根据数据集的表示形式,很容易分散传统更新规则的学习。在此基础上,本文提出了一套新的梯度更新规则,对训练数据集和学习参数的表示更为鲁棒。在各种数据集上进行的大量实验证实,提出的规则确实可以显着改善学习效果。此外,审查了高斯-伯努利RBM(GBRBM),它是RBM的变体,可以学习连续的实值数据集,并在其上测试了所建议的改进。实验表明,也可以对GBRBM进行改进。

著录项

  • 作者

    Cho KyungHyun;

  • 作者单位
  • 年度 2011
  • 总页数
  • 原文格式 PDF
  • 正文语种 en
  • 中图分类

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号