...
首页> 外文期刊>Behavioral Ecology and Sociobiology >Dealing with collinearity in behavioural and ecological data: model averaging and the problems of measurement error
【24h】

Dealing with collinearity in behavioural and ecological data: model averaging and the problems of measurement error

机译:处理行为和生态数据中的共线性:模型平均和测量误差问题

获取原文
获取原文并翻译 | 示例
           

摘要

There has been a great deal of recent discussion of the practice of regression analysis (or more generally, linear modelling) in behaviour and ecology. In this paper, I wish to highlight two factors that have been under-considered, collinearity and measurement error in predictors, as well as to consider what happens when both exist at the same time. I examine what the consequences are for conventional regression analysis (ordinary least squares, OLS) as well as model averaging methods, typified by information theoretic approaches based around Akaike's information criterion. Collinearity causes variance inflation of estimated slopes in OLS analysis, as is well known. In the presence of collinearity, model averaging reduces this variance for predictors with weak effects, but also can lead to parameter bias. When collinearity is strong or when all predictors have strong effects, model averaging relies heavily on the full model including all predictors and hence the results from this and OLS are essentially the same. I highlight that it is not safe to simply eliminate collinear variables without due consideration of their likely independent effects as this can lead to biases. Measurement error is also considered and I show that when collinearity exists, this can lead to extreme biases when predictors are collinear, have strong effects but differ in their degree of measurement error. I highlight techniques for dealing with and diagnosing these problems. These results reinforce that automated model selection techniques should not be relied on in the analysis of complex multivariable datasets.
机译:最近在行为和生态学方面对回归分析(或更一般地说,线性建模)的实践进行了大量讨论。在本文中,我希望强调两个尚未被充分考虑的因素,即预测变量中的共线性和测量误差,以及考虑当两者同时存在时会发生什么。我研究了常规回归分析(普通最小二乘法,OLS)以及模型平均方法的后果,这些方法以基于赤池信息准则的信息理论方法为代表。众所周知,共线性会导致OLS分析中估计斜率的方差膨胀。在存在共线性的情况下,模型平均会减少效果较弱的预测变量的方差,但也会导致参数偏差。当共线性很强或所有预测变量都具有强大影响时,模型平均在很大程度上依赖于包括所有预测变量的完整模型,因此,此结果与OLS的结果基本相同。我强调指出,简单地消除共线性变量而不充分考虑它们可能的独立影响是不安全的,因为这会导致偏差。还考虑了测量误差,并且我证明了当存在共线性时,当预测变量共线性时,这会导致极端偏差,影响很大,但其测量误差程度不同。我重点介绍了用于处理和诊断这些问题的技术。这些结果表明,在复杂的多变量数据集的分析中不应依赖自动模型选择技术。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号