【24h】

A Distributed Ensemble Scheme for nonlinear Support Vector Machine

机译:非线性支持向量机的分布式集成方案。

获取原文

摘要

We propose an ensemble scheme with a parallel computational structure which we call Distributed Ensemble Support Vector Machine (DESVM) to overcome the difficulties of large scale nonlinear Support Vector Machines (SVMs) in practice. The dataset is split into many stratified partitions. Each partition might be still too large to be solved by using conventional SVM solvers. We apply the reduced kernel trick to generate a nonlinear SVM classifier for each partition that can be treated as an approximation model based on the partial dataset. Then, we use a linear SVM classifier to fuse the nonlinear SVM classifiers that are generated from all data partitions. In this linear SVM training model, we treat each nonlinear SVM classifier as an “attribute” or an “expert”. In the ensemble phase, DESVM generates a fusion model which is a weighted combination of the nonlinear SVM classifiers. It can be explained as a weighted voting decision made by a group of experts. We test our proposed method on five benchmark datasets. The numerical results show that DESVM is competitive in accuracy and has a high speed-up. Thus, DESVM can be a powerful tool for binary classification problems with large scale not linearly separable datasets.
机译:我们提出了一种具有并行计算结构的集成方案,我们将其称为分布式集成支持向量机(DESVM),以克服实际中大规模非线性支持向量机(SVM)的困难。数据集分为许多分层分区。每个分区可能仍然太大,无法使用常规SVM求解器进行求解。我们应用简化的内核技巧为每个分区生成一个非线性SVM分类器,该分类器可以被视为基于部分数据集的近似模型。然后,我们使用线性SVM分类器融合从所有数据分区生成的非线性SVM分类器。在此线性SVM训练模型中,我们将每个非线性SVM分类器视为“属性”或“专家”。在集成阶段,DESVM生成融合模型,该模型是非线性SVM分类器的加权组合。可以解释为一组专家做出的加权投票决定。我们在五个基准数据集上测试了我们提出的方法。数值结果表明,DESVM在精度上具有竞争力,并且具有较高的加速比。因此,DESVM可以成为解决大规模不可线性分离数据集的二进制分类问题的有力工具。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号