...
首页> 外文期刊>Information Sciences: An International Journal >Scalable Gaussian Kernel Support Vector Machines with Sublinear Training Time Complexity
【24h】

Scalable Gaussian Kernel Support Vector Machines with Sublinear Training Time Complexity

机译:可扩展的高斯内核支持向量机,具有载于Sublinear培训时间复杂性

获取原文
获取原文并翻译 | 示例

摘要

Abstract Gaussian kernel Support Vector Machines (SVMs) deliver state-of-the-art generalization performance for non-linear classification, but the time complexity of their training process is at least quadratic w.r.t. the size of training set, preventing them from scaling up on large datasets. To address this issue, we propose a novel approach to large-scale kernel SVMs in sublinear time w.r.t. the size of training set, which combines three well-known and efficient techniques with theoretical guarantees. First, we subsample massive samples to reduce the sample size. Then, we use the random Fourier feature mapping on the subsamples to construct explicit random feature space where we can train a linear SVM to approximate the corresponding Gaussian kernel SVM. Finally, we use parallel algorithms to make our approach more scalable. Deriving the upper bounds of kernel matrix approximation error, hypothesis error and excess risk w.r.t. the size of training set and the dimension of random feature space, we establish the theoretical foundation of our proposed approach. In this way, we can reduce the time complexity of training kernel SVMs without sacrificing much accuracy. Our proposed approach achieves high accuracy and has sublinear training time complexity, which exhibits good scalability theoretically and empirically. ]]>
机译:<![cdata [ 抽象 高斯内核支持向量机(SVMS)为非线性分类提供最先进的泛化性能,但它们的时间复杂性培训过程至少是二次WRT培训集的大小,阻止它们在大型数据集中缩放。为了解决这个问题,我们提出了一种新的讲台时间W.R.T中的大规模内核SVM的方法。培训套装的大小,结合了三种具有理论保证的众所周知和有效的技术。首先,我们将大规模的样品分开以减少样品大小。然后,我们使用附带的随机傅里叶功能映射来构建显式随机特征空间,在其中我们可以培训线性SVM以近似相应的高斯内核SVM。最后,我们使用并行算法使我们的方法更加可扩展。导出内核矩阵近似误差,假设误差和过度风险W.r.t.的上限。培训集的规模和随机特征空间的维度,我们建立了我们提出的方法的理论基础。通过这种方式,我们可以减少培训内核SVM的时间复杂性而不会牺牲大量准确性。我们所提出的方法实现了高精度,并且具有高精度的培训时间复杂性,从理论上和经验上表现出良好的可扩展性。 ]]>

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号