首页> 外文会议>Machine learning and data mining in pattern recognition >A Large Margin Classifier with Additional Features
【24h】

A Large Margin Classifier with Additional Features

机译:具有附加功能的大型保证金分类器

获取原文
获取原文并翻译 | 示例

摘要

We consider the problem of learning classifiers from samples which have additional features that are absent due to noise or corruption of measurement. The common approach for handling missing features in discriminative models is first to complete their unknown values, and then a standard classification algorithm is employed over the completed data. In this paper, an algorithm which aims to maximize the margin of each sample in its own relevant subspace is proposed. We show how incomplete data can be classified directly without completing any missing features in a large-margin learning framework. Moreover, according to the theory of optimal kernel function, we proposed an optimal kernel function which is a convex composition of a set of linear kernel function to measure the similarity between additional features of each two samples. Based on the geometric interpretation of the margin, we formulate an objective function to maximize the margin of each sample in its own relevant subspace. In this formulation, we make use of the structural parameters trained from existing features and optimize the structural parameters trained from additional features only. A two-step iterative procedure for solving the objective function is proposed. By avoiding the pre-processing phase in which the data is completed, our algorithm could offer considerable computational saving. We demonstrate our results on a large number of standard benchmarks from UCI and the results show that our algorithm can achieve better or comparable classification accuracy compared to the existing algorithms.
机译:我们考虑从样本中学习分类器的问题,这些样本由于噪声或测量失真而缺少附加功能。在判别模型中处理缺失特征的常用方法是首先完成其未知值,然后对完成的数据采用标准分类算法。本文提出了一种旨在最大化每个样本在其自身相关子空间中的余量的算法。我们展示了如何在大型学习框架中不完成任何缺失功能的情况下直接分类不完整的数据。此外,根据最优核函数的理论,我们提出了一种最优核函数,它是一组线性核函数的凸组成,用于测量每两个样本的附加特征之间的相似性。基于边距的几何解释,我们制定了一个目标函数,以使每个样本在其自身相关子空间中的边距最大化。在此公式中,我们利用了从现有特征训练的结构参数,并且仅优化了从其他特征训练的结构参数。提出了求解目标函数的两步迭代过程。通过避免完成数据的预处理阶段,我们的算法可以节省大量计算量。我们在UCI的大量标准基准上证明了我们的结果,结果表明,与现有算法相比,我们的算法可以实现更好或相当的分类精度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号