首页> 外文学位 >Model-Based Methods for High-Dimensional Multivariate Analysis
【24h】

Model-Based Methods for High-Dimensional Multivariate Analysis

机译:高维多元分析的基于模型的方法

获取原文
获取原文并翻译 | 示例

摘要

This thesis consists of three main parts. In the first part, we propose a penalized likelihood method to fit the linear discriminant analysis model when the predictor is matrix valued. We simultaneously estimate the means and the precision matrix, which we assume has a Kronecker product decomposition. Our penalties encourage pairs of response category mean matrix estimators to have equal entries and also encourage zeros in the precision matrix estimator. To compute our estimators, we use a blockwise coordinate descent algorithm. To update the optimization variables corresponding to response category mean matrices, we use an alternating minimization algorithm that takes advantage of the Kronecker structure of the precision matrix. We show that our method can outperform relevant competitors in classification, even when our modeling assumptions are violated. We analyze an EEG dataset to demonstrate our method's interpretability and classification accuracy.;In the second part, we propose a class of estimators of the multivariate response linear regression coefficient matrix that exploits the assumption that the response and predictors have a joint multivariate normal distribution. This allows us to indirectly estimate the regression coefficient matrix through shrinkage estimation of the parameters of the inverse regression, or the conditional distribution of the predictors given the responses. We establish a convergence rate bound for estimators in our class and we study two examples, which respectively assume that the inverse regression's coefficient matrix is sparse and rank deficient. These estimators do not require that the forward regression coefficient matrix is sparse or has small Frobenius norm. Using simulation studies, we show that our estimators outperform competitors.;In the final part of this thesis, we propose a framework to shrink a user-specified characteristic of a precision matrix estimator that is needed to fit a predictive model. Estimators in our framework minimize the Gaussian negative log-likelihood plus an L1 penalty on a linear or affine function evaluated at the optimization variable corresponding to the precision matrix. We establish convergence rate bounds for these estimators and we propose an alternating direction method of multipliers algorithm for their computation. Our simulation studies show that our estimators can perform better than competitors when they are used to fit predictive models. In particular, we illustrate cases where our precision matrix estimators perform worse at estimating the population precision matrix while performing better at prediction.
机译:本论文主要包括三个部分。在第一部分中,我们提出了一种惩罚似然法,以在预测变量为矩阵值时适合线性判别分析模型。我们同时估计均值和精度矩阵,我们假设它们具有Kronecker积分解。我们的惩罚鼓励成对的响应类别均值矩阵估计值对具有相等的条目,并且还鼓励在精度矩阵估计值中使用零。为了计算我们的估计量,我们使用了逐块坐标下降算法。要更新与响应类别平均矩阵相对应的优化变量,我们使用交替最小化算法,该算法利用了精度矩阵的Kronecker结构。我们证明,即使违反了建模假设,我们的方法在分类上也能胜过相关竞争对手。我们分析了一个EEG数据集以证明我们方法的可解释性和分类准确性。在第二部分中,我们提出了多元响应线性回归系数矩阵的一类估计量,它利用了响应和预测变量具有联合多元正态分布的假设。这使我们能够通过收缩估计逆回归参数或给定响应的预测变量的条件分布来间接估计回归系数矩阵。我们为班级中的估计量建立一个收敛速率边界,并研究两个示例,分别假设逆回归系数矩阵稀疏且秩不足。这些估计量不需要正向回归系数矩阵稀疏或Frobenius范数较小。通过仿真研究,我们显示出我们的估计量优于竞争对手。;在本文的最后部分,我们提出了一个框架来缩小用户指定的精确矩阵估计量的特征,该特征需要适合预测模型。我们框架中的估计器将高斯负对数似然率加在与精度矩阵相对应的优化变量处评估的线性或仿射函数上的L1罚分最小化。我们为这些估计量建立了收敛速率边界,并提出了乘法器算法的交替方向方法进行计算。我们的仿真研究表明,当我们的估算器用于拟合预测模型时,它们的性能会优于竞争对手。特别是,我们说明了这样的情况:我们的精度矩阵估计器在估计总体精度矩阵时表现较差,而在预测方面却表现较好。

著录项

  • 作者

    Molstad, Aaron J.;

  • 作者单位

    University of Minnesota.;

  • 授予单位 University of Minnesota.;
  • 学科 Statistics.;Computer science.
  • 学位 Ph.D.
  • 年度 2017
  • 页码 102 p.
  • 总页数 102
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号