首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Adversarial Distillation of Bayesian Neural Network Posteriors
【24h】

Adversarial Distillation of Bayesian Neural Network Posteriors

机译:贝叶斯神经网络后验的对抗蒸馏

获取原文
           

摘要

Bayesian neural networks (BNNs) allow us to reason about uncertainty in a principled way. Stochastic Gradient Langevin Dynamics (SGLD) enables efficient BNN learning by drawing samples from the BNN posterior using mini-batches. However, SGLD and its extensions require storage of many copies of the model parameters, a potentially prohibitive cost, especially for large neural networks. We propose a framework, Adversarial Posterior Distillation, to distill the SGLD samples using a Generative Adversarial Network (GAN). At test-time, samples are generated by the GAN. We show that this distillation framework incurs no loss in performance on recent BNN applications including anomaly detection, active learning, and defense against adversarial attacks. By construction, our framework distills not only the Bayesian predictive distribution, but the posterior itself. This allows one to compute quantities such as the approximate model variance, which is useful in downstream tasks. To our knowledge, these are the first results applying MCMC-based BNNs to the aforementioned applications.
机译:贝叶斯神经网络(BNN)使我们能够以有原则的方式推理不确定性。随机梯度Langevin动力学(SGLD)通过使用迷你批从BNN后部提取样本来实现高效的BNN学习。但是,SGLD及其扩展要求存储许多模型参数副本,这是潜在的高昂成本,尤其是对于大型神经网络而言。我们提出了一个“对抗后蒸馏”框架,以使用“生成对抗网络”(GAN)蒸馏SGLD样本。在测试时,GAN会生成样本。我们表明,这种提炼框架不会对最近的BNN应用造成性能损失,包括异常检测,主动学习和对抗攻击。通过构造,我们的框架不仅提炼贝叶斯预测分布,而且提炼后验本身。这样一来,人们就可以计算诸如近似模型方差之类的数量,这对于下游任务非常有用。据我们所知,这是将基于MCMC的BNN应用于上述应用程序的第一个结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号