首页> 外文期刊>Journal of statistical computation and simulation >SUBSET SELECTION IN MULTIPLE LINEAR REGRESSION WITH HEAVY TAILED ERROR DISTRIBUTION
【24h】

SUBSET SELECTION IN MULTIPLE LINEAR REGRESSION WITH HEAVY TAILED ERROR DISTRIBUTION

机译:具有重尾误差分布的多线性回归中的子集选择

获取原文
获取原文并翻译 | 示例
           

摘要

Many subset selection methods in a multiple linear regression context such as Mallow's C_p-statistic are based on the Least squares (LS) estimator of regression coefficients. Whenever the distribution of the error variable is non-normal or heavy tailed, the LS estimator is known to perform 'poorly' and hence a method based on this estimator obviously selects a wrong subset. In this article, we propose a new criterion called N_p-criterion for subset selection in a multiple linear regression model whenever the distribution of error variable is heavy tailed. The new criterion is based on the nonparametric estimator of regression coefficients. The method is illustrated with examples simulated from a class of distributions such as laplace, Slash and mixture of normal distributions and it is demonstrated that in all these cases, it selects proper subsets while the classical subset selection methods such as C_p-statistic fail to do so. Further, it is shown that the proposed method is also useful whenever the data contains an influential or outlier observation.
机译:多重线性回归上下文中的许多子集选择方法(例如Mallow的C_p统计量)都是基于回归系数的最小二乘(LS)估计器。每当误差变量的分布为非正态分布或尾部很重时,已知LS估计量的执行效果很差,因此,基于该估计量的方法显然会选择错误的子集。在本文中,我们提出了一个新的准则N_p-criterion,用于只要误差变量的分布严重拖尾的多元线性回归模型中的子集选择。新标准基于回归系数的非参数估计量。举例说明了该方法,并从诸如laplace,Slash和正态分布的混合等类别的分布中模拟,并证明了在所有这些情况下,它都选择了适当的子集,而经典的子集选择方法(例如C_p-statistic)却没有做到所以。此外,表明,只要数据包含有影响或异常的观察结果,建议的方法也是有用的。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号