...
首页> 外文期刊>Statistics and computing >On the choice of the number of blocks with the incremental EM algorithm for the fitting of normal mixtures
【24h】

On the choice of the number of blocks with the incremental EM algorithm for the fitting of normal mixtures

机译:使用增量EM算法选择块数来拟合普通混合物

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

The EM algorithm is a popular method for parameter estimation in situations where the data can be viewed as being incomplete. As each E-step visits each data point on a given iteration, the EM algorithm requires considerable computation time in its application to large data sets. Two versions, the incremental EM (IEM) algorithm and a sparse version of the EM algorithm, were proposed recently by Neal R.M. and Hinton G.E. in Jordan M.I. (Ed.), Learning in Graphical Models, Kluwer, Dordrecht, 1998, pp. 355-368 to reduce the computational cost of applying the EM algorithm. With the IEM algorithm, the available n observations are divided into B (B ≤ n) blocks and the E-step is implemented for only a block of observations at a time before the next M-step is performed. With the sparse version of the EM algorithm for the fitting of mixture models, only those posterior probabilities of component membership of the mixture that are above a specified threshold are updated; the remaining component-posterior probabilities are held fixed. In this paper, simulations are performed to assess the relative performances of the IEM algorithm with various number of blocks and the standard EM algorithm. In particular, we propose a simple rule for choosing the number of blocks with the IEM algorithm. For the IEM algorithm in the extreme case of one observation per block, we provide efficient updating formulas, which avoid the direct calculation of the inverses and determinants of the component-covariance matrices. Moreover, a sparse version of the IEM algorithm (SPIEM) is formulated by combining the sparse E-step of the EM algorithm and the partial E-step of the IEM algorithm. This SPIEM algorithm can further reduce the computation time of the IEM algorithm.
机译:在可以将数据视为不完整的情况下,EM算法是一种流行的参数估计方法。当每个E步骤在给定迭代中访问每个数据点时,EM算法在应用于大型数据集时需要大量的计算时间。 Neal R.M最近提出了两种版本,即增量EM(IEM)算法和稀疏版本的EM算法。和欣顿G.E.在约旦(Ed。),《图形模型学习》,Kluwer,Dordrecht,1998年,第355-368页,以减少应用EM算法的计算成本。使用IEM算法,可用的n个观测值被分为B(B≤n)个块,并且在执行下一个M步之前,仅对一个观测值块执行E步。使用用于混合模型拟合的EM算法的稀疏版本,仅更新超过指定阈值的那些混合物的成员资格后验概率;其余分量后验概率保持固定。在本文中,进行了仿真以评估具有各种数量的块和标准EM算法的IEM算法的相对性能。特别是,我们提出了一个简单的规则,用于使用IEM算法选择块数。对于在每个块只有一个观测值的极端情况下的IEM算法,我们提供了有效的更新公式,该公式避免了直接计算分量协方差矩阵的逆和行列式。此外,通过组合EM算法的稀疏E步和IEM算法的部分E步来制定IEM算法(SPIEM)的稀疏版本。该SPIEM算法可以进一步减少IEM算法的计算时间。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号