首页> 外文会议>International Joint Conference on Neural Networks >An empirical analysis of different sparse penalties for autoencoder in unsupervised feature learning
【24h】

An empirical analysis of different sparse penalties for autoencoder in unsupervised feature learning

机译:无监督特征学习中自动编码器不同稀疏惩罚的实证分析

获取原文
获取外文期刊封面目录资料

摘要

Machine learning algorithms depend heavily on the data representation, which dominates its success in experiment accuracy. Autoencoder model structure is proposed to learn from data a good representation with the least possible amount of distortion. Furthermore, it has been proven that boosting sparsity when learning representation can significantly improve performance on classification tasks and also make the feature vector easy to interpret. One straightforward approach for autoencoder to obtain sparse representation is to impose sparse penalty on its overall cost function. Nevertheless, few comparative analysis has been conducted to evaluate which sparse penalty term works better. In this paper, we adopt L1 norm, L2 norm, Student-t penalties, which are rarely deployed to penalise the hidden unit outputs, and commonly used penalty KL-divergence in the literature. Then, we present a detailed analysis to evaluate which penalty achieves better result in terms of reconstruction error, sparseness of representation and classification performance on test datasets. Experimental study on MNIST, CIFAR-10, SVHN, OPTDIGITS and NORB datasets reveals that all these penalties achieve sparse representation and outperforms representations learned by pure autoencoder on classification performance and sparseness of feature vectors. Moreover, we hope this topics and the practices would provide insights for future research.
机译:机器学习算法在很大程度上取决于数据表示形式,这决定了它在实验准确性方面的成功。提出了自动编码器模型结构,以从数据中学习尽可能少的失真的良好表示。此外,已经证明,在学习表示时提高稀疏度可以显着提高分类任务的性能,并使特征向量易于解释。自动编码器获得稀疏表示的一种直接方法是对其总体成本函数施加稀疏惩罚。但是,很少进行比较分析来评估哪个稀疏惩罚项效果更好。在本文中,我们采用L1范数,L2范数,Student-t罚分,它们很少用于惩罚隐藏的单位输出,并且在文献中通常采用罚分KL散度。然后,我们进行了详细的分析,以评估在重构误差,表示的稀疏性和测试数据集分类性能方面哪些惩罚可以取得更好的结果。对MNIST,CIFAR-10,SVHN,OPTDIGITS和NORB数据集的实验研究表明,所有这些惩罚都实现了稀疏表示,并且优于纯自动编码器在分类性能和特征向量稀疏性方面的学习。此外,我们希望这个主题和实践能够为将来的研究提供见识。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号