首页> 外文期刊>Journal of machine learning research >Model Selection Through Sparse Maximum Likelihood Estimation for Multivariate Gaussian or Binary Data
【24h】

Model Selection Through Sparse Maximum Likelihood Estimation for Multivariate Gaussian or Binary Data

机译:多元高斯或二进制数据的稀疏最大似然估计模型选择

获取原文
           

摘要

We consider the problem of estimating the parameters of a Gaussian orbinary distribution in such a way that the resulting undirectedgraphical model is sparse. Our approach is to solve a maximumlikelihood problem with an added l1-norm penalty term. Theproblem as formulated is convex but the memory requirements andcomplexity of existing interior point methods are prohibitive forproblems with more than tens of nodes. We present two new algorithmsfor solving problems with at least a thousand nodes in the Gaussiancase. Our first algorithm uses block coordinate descent, and can beinterpreted as recursive l1-norm penalized regression. Oursecond algorithm, based on Nesterov's first order method, yields acomplexity estimate with a better dependence on problem size thanexisting interior point methods. Using a log determinant relaxationof the log partition function (Wainwright and Jordan, 2006), we show that thesesame algorithms can be used to solve an approximate sparse maximumlikelihood problem for the binary case. We test our algorithms onsynthetic data, as well as on gene expression and senate votingrecords data. color="gray">
机译:我们考虑以这样一种方式来估计高斯二元分布参数的问题,即所得的无向图模型是稀疏的。我们的方法是通过添加 l 1 -范数惩罚项来解决最大似然问题。提出的问题是凸的,但是现有内点方法的内存需求和复杂性对于节点数超过数十的问题是不允许的。我们提出了两种新的算法来解决高斯案例中至少一千个节点的问题。我们的第一个算法使用块坐标下降,并且可以解释为递归 l 1 -范数罚分式回归。我们的第二种算法基于Nesterov的一阶方法,得出的复杂度估计比现有的内部点方法对问题大小的依赖性更好。使用对数确定函数放宽对数分区函数(Wainwright和Jordan,2006),我们证明了这些相同的算法可用于解决二进制情况下的近似稀疏最大似然问题。我们在合成数据以及基因表达和参议院投票记录数据上测试我们的算法。 color =“ gray”>

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号