首页> 外文会议>International conference on neural information processing >Learning Deep Neural Network Based Kernel Functions for Small Sample Size Classification
【24h】

Learning Deep Neural Network Based Kernel Functions for Small Sample Size Classification

机译:学习基于深度神经网络的核函数进行小样本量分类

获取原文

摘要

Kernel learning is to learn a kernel function based on the set of all sample pairs from training data. Even for small sample size classification tasks, the set size is mostly large enough to make a complex kernel that holds lots of parameters being well optimized. Hence, the complex kernel can be helpful in improving classification performance via providing more meaningful feature representation in kernel induced feature space. In this paper, we propose to embed a deep neural network (DNN) into kernel functions, taking its output as kernel parameter to adjust the feature representations adaptively. Two kind of DNN based kernels are denned, and both of them are proved to satisfy the Mercer theorem. Considering the connection between kernel and classifier, we optimize the proposed DNN based kernels by exploiting the GMKL alternating optimization framework. A stochastic gradient descent (SGD) based algorithm is also proposed, which still implements alternating optimization in each iteration. Furthermore, an incremental batch size method is given to reduce gradient noise gradually in optimization process. Experimental results show that our method performed better than the typical methods.
机译:内核学习是根据训练数据中所有样本对的集合来学习内核功能。即使对于小样本大小的分类任务,集合大小也大到足以使包含许多参数的复杂内核得到很好的优化。因此,复杂内核可以通过在内核诱导的特征空间中提供更有意义的特征表示来帮助提高分类性能。在本文中,我们建议将深度神经网络(DNN)嵌入到内核函数中,并以其输出作为内核参数来自适应地调整特征表示。定义了两种基于DNN的内核,并证明它们都满足Mercer定理。考虑到内核和分类器之间的联系,我们通过利用GMKL交替优化框架来优化建议的基于DNN的内核。还提出了一种基于随机梯度下降(SGD)的算法,该算法仍在每次迭代中实现交替优化。此外,在优化过程中,采用增量批处理方法逐步降低梯度噪声。实验结果表明,我们的方法比典型方法具有更好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号