【24h】

Generalization Error of Linear Neural Networks in Unidentifiable Cases

机译:不可识别情况下线性神经网络的泛化误差

获取原文
获取原文并翻译 | 示例

摘要

The statistical asymptotic theory is often used in theoretical results in computational and statistical learning theory. It describes the limiting distribution of the maximum likelihood estimator (MLE) as an normal distribution. However, in layered models such as neural networks, the regularity condition of the asymptotic theory is not necessarily satisfied. the true parameter is not identifiable, if the target function can be realized by a network of smaller size than the size of the model. There has been little known on hte behavior of the MLE in these cases of neural networks. In this paper, we analyze the expectation of the generalization error of three-layer linear neural networks, and elucidate a strange behavior in unidentifiable cases. We show that the expectation of the generalization error in the unidentifiable cases is larger than what is given by the usual asymptotic theory, and dependent on the rank of the target function.
机译:统计渐近理论经常用于计算和统计学习理论的理论结果中。它将最大似然估计器(MLE)的极限分布描述为正态分布。但是,在诸如神经网络之类的分层模型中,渐进理论的正则性条件不一定得到满足。如果目标功能可以通过比模型大小小的网络来实现,则无法确定真实参数。在这些神经网络的情况下,关于MLE的行为还鲜为人知。在本文中,我们分析了三层线性神经网络的泛化误差的预期,并阐明了在无法识别的情况下的奇怪行为。我们表明,在无法确定的情况下,对泛化误差的期望大于通常的渐近理论所给出的期望,并且取决于目标函数的等级。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号