...
首页> 外文期刊>IEEE Transactions on Pattern Analysis and Machine Intelligence >Fast SVM training algorithm with decomposition on very large data sets
【24h】

Fast SVM training algorithm with decomposition on very large data sets

机译:快速SVM训练算法,可分解非常大的数据集

获取原文
获取原文并翻译 | 示例

摘要

Training a support vector machine on a data set of huge size with thousands of classes is a challenging problem. This paper proposes an efficient algorithm to solve this problem. The key idea is to introduce a parallel optimization step to quickly remove most of the nonsupport vectors, where block diagonal matrices are used to approximate the original kernel matrix so that the original problem can be split into hundreds of subproblems which can be solved more efficiently. In addition, some effective strategies such as kernel caching and efficient computation of kernel matrix are integrated to speed up the training process. Our analysis of the proposed algorithm shows that its time complexity grows linearly with the number of classes and size of the data set. In the experiments, many appealing properties of the proposed algorithm have been investigated and the results show that the proposed algorithm has a much better scaling capability than Libsvm, SVM/sup light/, and SVMTorch. Moreover, the good generalization performances on several large databases have also been achieved.
机译:在具有数千个类的庞大数据集上训练支持向量机是一个具有挑战性的问题。本文提出了一种有效的算法来解决这个问题。关键思想是引入并行优化步骤,以快速删除大多数非支持向量,其中使用块对角矩阵近似原始内核矩阵,以便可以将原始问题分解为数百个子问题,可以更有效地解决这些问题。此外,还集成了一些有效的策略(例如内核缓存和内核矩阵的有效计算)以加快训练过程。我们对提出的算法的分析表明,其时间复杂度随类数和数据集大小线性增长。在实验中,研究了该算法的许多吸引人的性质,结果表明,与Libsvm,SVM / sup light /和SVMTorch相比,该算法具有更好的缩放能力。此外,还已经在几个大型数据库上实现了良好的泛化性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号