首页> 外文期刊>Machine Learning >Weak approximation of transformed stochastic gradient MCMC
【24h】

Weak approximation of transformed stochastic gradient MCMC

机译:变换随机梯度MCMC的弱近似

获取原文
获取原文并翻译 | 示例
           

摘要

Stochastic gradient Langevin dynamics (SGLD) is a computationally efficient sampler for Bayesian posterior inference given a large scale dataset and a complex model. Although SGLD is designed for unbounded random variables, practical models often incorporate variables within a bounded domain, such as non-negative or a finite interval. The use of variable transformation is a typical way to handle such a bounded variable. This paper reveals that several mapping approaches commonly used in the literature produce erroneous samples from theoretical and empirical perspectives. We show that the change of random variable in discretization using an invertible Lipschitz mapping function overcomes the pitfall as well as attains the weak convergence, while the other methods are numerically unstable or cannot be justified theoretically. Experiments demonstrate its efficacy for widely-used models with bounded latent variables, including Bayesian non-negative matrix factorization and binary neural networks.
机译:随机梯度Langevin Dynamics(SGLD)是给定大型数据集和复杂模型的贝叶斯后部推理的计算上有效的采样器。尽管SGLD被设计用于无限的随机变量,但实际模型通常在有界域内的变量融入,例如非负或有限间隔。使用变换是处理这种有界变量的典型方法。本文揭示了文献中常用的几种映射方法从理论和实证角度产生错误的样本。我们表明,使用可逆的Lipschitz映射函数的离散化随机变量的变化克服了缺陷以及达到弱收敛,而其他方法在实际上是不稳定的,或者理论上不能合理。实验证明了具有界限潜变量的广泛使用模型的功效,包括贝叶斯非负矩阵分解和二元神经网络。

著录项

  • 来源
    《Machine Learning》 |2020年第10期|1903-1923|共21页
  • 作者单位

    Univ Tokyo Grad Sch Frontier Sci Dept Complex Sci & Engn 5-1-5 Kashiwanoha Kashiwa Chiba 2778561 Japan|RIKEN Chuo Ku 1-4-1 Nihonbashi Tokyo 1030027 Japan;

    NTT Corp NTT Commun Sci Labs 2-4 Hikaridai Seika Cho Kyoto 6190237 Japan;

    RIKEN Chuo Ku 1-4-1 Nihonbashi Tokyo 1030027 Japan|Univ Tokyo Grad Sch Informat Sci & Technol Dept Comp Sci Bunkyo Ku 7-3-1 Hongo Tokyo 1130033 Japan;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Stochastic gradient MCMC; Transform; Convergence analysis; Ito process;

    机译:随机梯度MCMC;变换;收敛分析;ITO过程;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号