【24h】

Incremental Mixture Learning for Clustering Discrete Data

机译:用于离散数据聚类的增量混合学习

获取原文
获取原文并翻译 | 示例

摘要

This paper elaborates on an efficient approach for clustering discrete data by incrementally building multinomial mixture models through likelihood maximization using the Expectation-Maximization (EM) algorithm. The method adds sequentially at each step a new multinomial component to a mixture model based on a combined scheme of global and local search in order to deal with the initialization problem of the EM algorithm. In the global search phase several initial values are examined for the parameters of the multinomial component. These values are selected from an appropriately defined set of initialization candidates. Two methods are proposed here to specify the elements of this set based on the agglomerative and the kd-tree clustering algorithms. We investigate the performance of the incremental learning technique on a synthetic and a real dataset and also provide comparative results with the standard EM-based multinomial mixture model.
机译:本文阐述了一种有效的方法,该方法通过使用期望最大化(EM)算法通过似然最大化来逐步建立多项式混合模型,从而对离散数据进行聚类。该方法在每个步骤中基于全局和局部搜索的组合方案将新的多项式分量顺序地添加到混合模型中,以处理EM算法的初始化问题。在全局搜索阶段,将检查几个初始值以获取多项式分量的参数。这些值是从一组适当定义的初始化候选中选择的。本文提出了两种方法,用于基于凝聚和kd-tree聚类算法指定该集合的元素。我们研究了在合成数据集和真实数据集上的增量学习技术的性能,并且还提供了基于标准基于EM的多项式混合模型的比较结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号