【24h】

Training Structural SVMs with Kernels Using Sampled Cuts

机译:使用采样切口使用核训练结构SVM

获取原文

摘要

Discriminative training for structured outputs has found increasing applications in areas such as natural language processing, bioinformatics, information retrieval, and computer vision. Focusing on large-margin methods, the most general (in terms of loss function and model structure) training algorithms known to date are based on cutting-plane approaches. While these algorithms are very efficient for linear models, their training complexity becomes quadratic in the number of examples when kernels are used. To overcome this bottleneck, we propose new training algorithms that use approximate cutting planes and random sampling to enable efficient training with kernels. We prove that these algorithms have improved time complexity while providing approximation guarantees. In empirical evaluations, our algorithms produced solutions with training and test error rates close to those of exact solvers. Even on binary classification problems where highly optimized conventional training methods exist (e.g. SVM-light), our methods are about an order of magnitude faster than conventional training methods on large datasets, while remaining competitive in speed on datasets of medium size.
机译:发现对结构化输出的歧视性培训在自然语言处理,生物信息学,信息检索和计算机视觉等领域中的应用越来越广泛。着眼于大幅度的方法,迄今为止已知的最通用的(就损失函数和模型结构而言)训练算法都是基于切面方法。尽管这些算法对于线性模型非常有效,但是当使用内核时,它们的训练复杂度在示例数上变成了平方。为了克服这个瓶颈,我们提出了新的训练算法,该算法使用近似的切割平面和随机采样来实现内核的有效训练。我们证明了这些算法在提供逼近保证的同时改善了时间复杂度。在经验评估中,我们的算法产生的解决方案的训练和测试错误率接近于精确求解器。即使在存在高度优化的常规训练方法(例如SVM-light)的二进制分类问题上,我们的方法也比大型数据集上的常规训练方法快一个数量级,同时在中等大小的数据集上仍具有竞争力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号