...
首页> 外文期刊>IEEE Transactions on Neural Networks >A Kernel-Based Two-Class Classifier for Imbalanced Data Sets
【24h】

A Kernel-Based Two-Class Classifier for Imbalanced Data Sets

机译:基于内核的两类分类器,用于不平衡数据集

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Many kernel classifier construction algorithms adopt classification accuracy as performance metrics in model evaluation. Moreover, equal weighting is often applied to each data sample in parameter estimation. These modeling practices often become problematic if the data sets are imbalanced. We present a kernel classifier construction algorithm using orthogonal forward selection (OFS) in order to optimize the model generalization for imbalanced two-class data sets. This kernel classifier identification algorithm is based on a new regularized orthogonal weighted least squares (ROWLS) estimator and the model selection criterion of maximal leave-one-out area under curve (LOO-AUC) of the receiver operating characteristics (ROCs). It is shown that, owing to the orthogonalization procedure, the LOO-AUC can be calculated via an analytic formula based on the new regularized orthogonal weighted least squares parameter estimator, without actually splitting the estimation data set. The proposed algorithm can achieve minimal computational expense via a set of forward recursive updating formula in searching model terms with maximal incremental LOO-AUC value. Numerical examples are used to demonstrate the efficacy of the algorithm
机译:许多内核分类器构造算法在模型评估中采用分类精度作为性能指标。此外,在参数估计中通常将相等的加权应用于每个数据样本。如果数据集不平衡,则这些建模实践通常会出现问题。我们提出一种使用正交前向选择(OFS)的内核分类器构造算法,以优化不平衡两类数据集的模型概括。该内核分类器识别算法基于新的正则化的正交加权最小二乘(ROWLS)估计器和接收器工作特性(ROC)的曲线下最大留一留面积模型选择标准(LOO-AUC)。结果表明,由于采用了正交化程序,因此可以基于解析的公式基于新的正则化的正交加权最小二乘参数估计器来计算LOO-AUC,而无需实际拆分估计数据集。在搜索具有最大增量LOO-AUC值的模型项时,该算法可以通过一组前向递归更新公式来实现最小的计算开销。数值例子证明了该算法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号