【24h】

Learning by /spl alpha/-divergence

机译:通过/ spl alpha / -divergence学习

获取原文

摘要

In the present paper, we propose a new cost function, called /spl alpha/-divergence, which is a generalized version of the relative entropy or the Kullback's divergence measure in neural network. The most fundamental characteristics of this /spl alpha/-divergence are summarized by the following three points: 1) by changing the parameter /spl alpha/ for the /spl alpha/-divergence, multiple cost functions can be obtained to be used for different purposes or problems; 2) /spl alpha/-divergence is effective in direct proportion to the error between targets and outputs, eliminating the derivative of the sigmoidal function; and 3) the /spl alpha/-divergence has an effect on eliminating saturated units. We formulated an update rule to minimize /spl alpha/-divergence, and applied the method to the acquisition of the grammatical competence. Experimental results confirmed marked improvement in the generalization by using /spl alpha/-divergence. This improvement is due to the property of /spl alpha/-divergence whose derivative is effective especially for eliminating saturated units.
机译:在本文中,我们提出了一个新的成本函数/ spl alpha / -divergence,它是神经网络中相对熵或Kullback散度测度的广义形式。 / spl alpha / -divergence的最基本特征概括为以下三点:1)通过更改/ spl alpha / -divergence的参数/ spl alpha /,可以获取多个成本函数以用于不同的目的或问题; 2)/ spl alpha /-散度与目标和输出之间的误差成正比,有效,消除了S形函数的导数; 3)/ spl alpha / -divergence有消除饱和单元的作用。我们制定了一个更新规则以最小化/ spl alpha /-发散,并将该方法应用于语法能力的获得。实验结果证实,通过使用/ spl alpha / -divergence,泛化效果显着提高。该改进归因于/ spl alpha / -divergence的属性,其导数尤其对于消除饱和单元有效。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号