首页> 外文期刊>JMLR: Workshop and Conference Proceedings >On the Dynamics of Gradient Descent for Autoencoders
【24h】

On the Dynamics of Gradient Descent for Autoencoders

机译:关于自动化器梯度下降的动态

获取原文
       

摘要

We provide a series of results for unsupervised learning with autoencoders. Specifically, we study shallow two-layer autoencoder architectures with shared weights. We focus on three generative models for data that are common in statistical machine learning: (i) the mixture-of-gaussians model, (ii) the sparse coding model, and (iii) the sparsity model with non-negative coefficients. For each of these models, we prove that under suitable choices of hyperparameters, architectures, and initialization, autoencoders learned by gradient descent can successfully recover the parameters of the corresponding model. To our knowledge, this is the first result that rigorously studies the dynamics of gradient descent for weight-sharing autoencoders. Our analysis can be viewed as theoretical evidence that shallow autoencoder modules indeed can be used as feature learning mechanisms for a variety of data models, and may shed insight on how to train larger stacked architectures with autoencoders as basic building blocks.
机译:我们为与AutoEncoders无监督学习提供了一系列结果。具体而言,我们研究具有共享权重的浅两层AutoEncoder架构。我们专注于三个生成模型,用于统计机器学习中常见的数据:(i)稀疏编码模型(ii)稀疏编码模型,(iii)具有非负系数的稀疏模型。对于这些模型中的每一个,我们证明,在具有梯度下降的合适选择的超参数,架构和初始化的基础上,可以成功恢复相应模型的参数。为我们的知识,这是第一个严格研究重量共享自身额度梯度下降的动态的结果。我们的分析可以被视为理论上证据,即浅宇的AutoEncoder模块确实可以用作各种数据模型的特征学习机制,并且可以阐述如何使用AutoEncoders作为基本构建块培训更大的堆叠体系结构。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号