首页> 中文期刊>计算机与现代化 >一种启发式线性回归损失函数选取方法

一种启发式线性回归损失函数选取方法

     

摘要

损失函数度量回归分析中,信息损失和错误的程度是机器学习算法最小化的目标函数.本文研究在有限数据集上线性回归分析的损失函数选取方法.对于给定的噪声密度,存在一个满足一致性条件的最优损失函数(如噪声密度满足高斯分布,则常见的最优损失函数是平方损失函数).但在实际应用中,噪声密度往往是不确定的,且训练样本集有限.一些统计信息可用来对有限信息环境下的损失函数进行选取,但这些统计信息是基于一些一致性假设且在有限的样本集上不一定有效.针对这些问题,借鉴Vapnik的ε-insensitive损失函数,提出一种启发式的基于样本数目及噪声方差的参数设置方法.实验结果表明,与常用的平方损失函数及Huber的least-modulus loss相比,本文的损失函数性能更健壮且预测效率更准确.%Loss function is used to quantify information loss and false degree in regression analysis.This paper addresses heuristic loss function selection for linear regression.For a given noise density, there exists an optimal loss function under an asymptotic setting i.e.squared loss is optimal for Gaussian noise density.However, in real-life applications the noise density is always unknown and the training samples are finite.Robust statistics provides ways for selecting the loss function using statistical information about noise density, however robust statistics is based on asymptotic assumption and may not be well applied for finite sample data sets.For such practical problems, we try to utilize concept of Vapnik''s ε-insensitive loss function.We propose a heuristic method for setting the value of ε as a function of samples and noise variance.Experimental comparisons for linear regression problems show that the proposed loss function performs more robustly performance and yields higher prediction accuracy compared with popular squared loss and Huber''s least-modulus loss.

著录项

相似文献

  • 中文文献
  • 外文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号