首页> 外文学位 >Penalized Regression Methods with Application to Domain Selection and Outlier Detection.
【24h】

Penalized Regression Methods with Application to Domain Selection and Outlier Detection.

机译:惩罚回归方法及其在域选择和离群值检测中的应用。

获取原文
获取原文并翻译 | 示例

摘要

Variable selection is one of the most important problems in statistical literature, and a popular method is the penalized regression. In this dissertation, we investigate two types of variable selection problems via penalized regression. The first problem is related to the domain selection for the varying coefficient model, which identifies important regions of the varying coefficient that are related to the response. The second problem is related to variable selection problem for the linear model with the existence of outliers, and deal with variable selection and outlier detection simultaneously.;In Chapter 1, we give some introduction and background by overview of the variable selection methods, the local polynomial regression and the robust regression. In Chapter 2, we consider the varying coefficient model which allows the relationship between the predictors and response to vary across the domain of interest, such as time. In applications, it is possible that certain predictors only affect the response in particular regions and not everywhere. This corresponds to identifying the domain where the varying coefficient is nonzero. Towards this goal, we incorporate local polynomial smoothing and penalized regression into one framework. We establish asymptotic properties of our penalized estimators and show that they enjoy the oracle properties in the sense that they have the same bias and asymptotic variance as the local polynomial estimators as if the sparsity is known as a priori. The choice of appropriate bandwidth and computational algorithms are discussed. In Chapter 2, we study the outlier detection and variable selection problem in linear regression. A mean shift parameter is added to the linear model to reflect the effect of the outlier, where an outlier has a nonzero shift parameter. We then apply an adaptive regularization on these shift parameters and shrink most of them to zero. For those observations with nonzero mean shift parameter estimates, they are regarded as outliers. Meanwhile, an L1 penalty is added to the regression parameters to select important predictors. We propose an efficient algorithm to solve this jointly penalized optimization problem and use the extended BIC tuning method to select the regularization parameters since the number of parameters exceeds the sample size. Theoretical results are provided in terms of high breakdown point, full efficiency as well as outlier detection consistency. We illustrate our method with simulation and real data. Our method is extended to high-dimensional problems with p &
机译:变量选择是统计文献中最重要的问题之一,一种流行的方法是惩罚回归。本文通过惩罚回归研究了两种类型的变量选择问题。第一个问题与可变系数模型的域选择有关,该域选择确定了与响应相关的可变系数的重要区域。第二个问题与存在离群值的线性模型的变量选择问题有关,并同时处理变量选择和离群值检测。在第一章中,我们概述了变量选择方法,介绍了局部背景和局部背景。多项式回归和稳健回归。在第2章中,我们考虑了可变系数模型,该模型允许预测变量与响应之间的关系在感兴趣的领域(例如时间)上变化。在应用程序中,某些预测变量可能仅影响特定区域而不是任何地方的响应。这对应于识别变化系数非零的域。为了实现这一目标,我们将局部多项式平滑和惩罚回归合并到一个框架中。我们建立了惩罚估计量的渐近性质,并表明它们在与局部多项式估计量具有相同的偏差和渐近方差的意义上享受甲骨文的性质,就好像稀疏性被称为先验值一样。讨论了适当带宽的选择和计算算法。在第二章中,我们研究线性回归中的离群值检测和变量选择问题。将平均移位参数添加到线性模型以反映异常值的影响,其中异常值具有非零的移位参数。然后,我们对这些移位参数应用自适应正则化,并将其中大多数缩小为零。对于那些非零均值漂移参数估计值的观测,它们被视为离群值。同时,将L1罚分添加到回归参数以选择重要的预测变量。我们提出了一种有效的算法来解决该联合惩罚优化问题,并使用扩展的BIC调整方法来选择正则化参数,因为参数数量超过了样本大小。从高击穿点,充分的效率以及离群的检测一致性方面提供了理论结果。我们用仿真和真实数据说明了我们的方法。我们的方法扩展到p&

著录项

  • 作者

    Kong, Dehan.;

  • 作者单位

    North Carolina State University.;

  • 授予单位 North Carolina State University.;
  • 学科 Statistics.
  • 学位 Ph.D.
  • 年度 2013
  • 页码 70 p.
  • 总页数 70
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号