...
首页> 外文期刊>WSEAS Transactions on Computers >Comparison of Expectation Maximization and K-means Clustering Algorithms with Ensemble Classifier Model
【24h】

Comparison of Expectation Maximization and K-means Clustering Algorithms with Ensemble Classifier Model

机译:与合奏分类器模型的期望最大化与K-meast聚类算法的比较

获取原文
获取原文并翻译 | 示例
           

摘要

In data mining, classification learning is broadly categorized into two categories; supervised and unsupervised. In the former category, the training example is learned and the hidden class is predicted to represent the appropriate class. The class is known, but it is hidden from the learning model. Unlike supervised, unsupervised directly build the learning model for unlabeled example. Clustering is one of the means in data mining of predicting the class based on separating the data categories from similar features. Expectation maximization (EM) is one of the representatives clustering algorithms which have broadly applied in solving classification problems by improving the density of data using the probability density function. Meanwhile, Kmeans clustering algorithm has also been reported has widely known for solving most unsupervised classification problems. Unlike EM, K-means performs the clustering by measuring the distance between the data centroid and the object within the same cluster. On top of that, random forest ensemble classifier model has reported successive perform in most classification and pattern recognition problems. The expanding of randomness layer in the traditional decision tree is able to increase the diversity of classification accuracy. However, the combination of clustering and classification algorithm might rarely be explored, particularly in the context of an ensemble classifier model. Furthermore, the classification using original attribute might not guarantee to achieve high accuracy. In such states, it could be possible some of the attributes might overlap or may redundant and also might incorrectly place in its particular cluster. Hence, this situation is believed in yielding of decreasing the classification accuracy. In this article, we present the exploration on the combination of the clustering based algorithm with an ensemble classification learning. EM and K-means clustering algorithms are used to cluster the multi-class classification attribute according to its relevance criteria and afterward, the clustered attributes are classified using an ensemble random forest classifier model. In our experimental analysis, ten widely used datasets from UCI Machine Learning Repository and additional two accelerometer human activity recognition datasets are utilized.
机译:在数据挖掘中,分类学习大致分为两类;监督和无人监督。在前类别中,学习训练示例,预测隐藏类别表示相应的类。课堂是已知的,但它隐藏在学习模型中。与监督不同,无监督直接构建未标记示例的学习模型。群集是基于与类似特征的数据类别分离的数据挖掘数据挖掘的方法之一。期望最大化(EM)是通过使用概率密度函数提高数据密度来求解分类问题的代表聚类算法之一。同时,据报道,奎氏群组集群算法已普遍普遍地众所周知,以解决大多数无监督的分类问题。与EM不同,K-means通过测量数据质心和同一群集中的对象之间的距离来执行群集。在此之上,随机森林集合分类器模型在大多数分类和模式识别问题中报告了连续执行。传统决策树中的随机层的扩展能够增加分类准确性的分集。然而,可以很少探索聚类和分类算法的组合,特别是在集合分类器模型的上下文中。此外,使用原始属性的分类可能无法保证实现高精度。在这样的状态中,可以将一些属性重叠或可能冗余,并且也可能在其特定群集中不正确地放置。因此,据信这种情况促成降低分类准确性。在本文中,我们展示了基于集群基于集群的算法与集群分类学习的探索。 EM和K-Means群集算法用于根据其相关性标准群集多级分类属性,然后使用集群属性使用集群属性使用集群随机林分类器模型进行分类。在我们的实验分析中,利用了来自UCI机器学习存储库的十种广泛使用的数据集和额外的两个加速度计人类活动识别数据集。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号