首页> 外文会议>SPIE Conference on Computer-Aided Diagnosis >Neural Network Training by Maximization of the Area Under the ROC Curve: Application to Characterization of Masses on Breast Ultrasound as Malignant or Benign
【24h】

Neural Network Training by Maximization of the Area Under the ROC Curve: Application to Characterization of Masses on Breast Ultrasound as Malignant or Benign

机译:通过ROC曲线下的区域的最大化培训神经网络培训:应用于乳房超声的群体表征为恶性或良性

获取原文
获取外文期刊封面目录资料

摘要

Back-propagation neural networks (BPNs) are traditionally trained using error measures such as sum-of-squares or cross-entropy. If the training sample size is small, and the neural network has a large number of hidden layer nodes, the BPN may be overtrained, i.e., it may fit the training data well, but may generalize poorly to independent test data. In this study, we investigated a training technique that maximized the approximate area under the ROC curve (AUC) to reduce overtraining. In general, the non-parametric AUC is a discontinuous and non-differentiable function of the neural network output, which makes it unsuitable for gradient descent algorithms such as back-propagation. We used a semidifferentiable approximation to AUC, which appeared to provide reasonable training for the data sets explored in this study. We performed a simulation study using synthetic data sets consisting of Gaussian mixtures to investigate the behavior of this new technique with respect to overtraining. Our results indicated that an artificial neural network trained using the AUC-maximization method is less prone to overtraining. The advantage of the AUC-maximization method was consistently observed over different values of hidden layer BPN nodes, training sample sizes, and the dimensionality of the feature spaces evaluated in our simulation study. For a five-hidden-node BPN trained using 50 training samples per class, the average test AUC was 0.896 (standard deviation (SD): 0.026) with AUC-maximization and 0.856 (SD: 0.028) with the sum-of-squares method. The gain in test performance by the AUC-maximization method over the traditional BPN training was greater when the training sample size was smaller. We also applied this new method to a data set previously acquired for characterization of masses on breast ultrasound as malignant or benign. Our results with this real-world data set had the same trend as with our simulation data sets in that the AUC-maximization technique was less prone to overtraining than the sum-of-squares method.
机译:背部传播神经网络(BPN)传统上使用诸如平方和跨熵的误差测量训练。如果训练样本大小很小,并且神经网络具有大量的隐藏层节点,则可以过度训练BPN,即,它可以良好地拟合训练数据,但可能概括为独立的测试数据。在这项研究中,我们调查了一种训练技术,最大化ROC曲线(AUC)下的近似面积,以减少过度训练。通常,非参数AUC是神经网络输出的不连续和不可微差的函数,这使得它不适用于诸如反向传播的梯度下降算法。我们对AUC使用了一个半异端近似,似乎为本研究探索的数据集提供了合理的培训。我们使用由高斯混合组成的合成数据集进行仿真研究,以研究这种新技术的行为持续训练。我们的结果表明,使用AUC-Maximization方法培训的人工神经网络不太容易过度训练。在我们的仿真研究中评估的特征空间的不同值,始终观察到AUC最大化方法的优点是在不同的隐藏层BPN节点,训练样本尺寸的值中观察到。对于使用每类50次训练样本培训的五个隐藏节点BPN,平均测试AUC为0.896(标准差(SD):0.026),具有AUC最大化和0.856(SD:0.028),具有方块方法。当训练样本尺寸较小时,通过传统的BPN训练的AUC最大化方法对测试性能的增益更大。我们还将这种新方法应用于先前获取的数据集,以便在乳房超声为恶性或良性地区表征群众。我们使用这种真实数据集的结果与我们的仿真数据集具有相同的趋势,因为AUC最大化技术不太容易超过超支比率方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号