首页> 外文会议>IEEE Symposium Series on Computational Intelligence >X-CNN: Cross-modal convolutional neural networks for sparse datasets
【24h】

X-CNN: Cross-modal convolutional neural networks for sparse datasets

机译:X-CNN:稀疏数据集的跨模态卷积神经网络

获取原文

摘要

In this paper we propose cross-modal convolutional neural networks (X-CNNs), a novel biologically inspired type of CNN architectures, treating gradient descent-specialised CNNs as individual units of processing in a larger-scale network topology, while allowing for unconstrained information flow and/or weight sharing between analogous hidden layers of the network-thus generalising the already well-established concept of neural network ensembles (where information typically may flow only between the output layers of the individual networks). The constituent networks are individually designed to learn the output function on their own subset of the input data, after which cross-connections between them are introduced after each pooling operation to periodically allow for information exchange between them. This injection of knowledge into a model (by prior partition of the input data through domain knowledge or unsupervised methods) is expected to yield greatest returns in sparse data environments, which are typically less suitable for training CNNs. For evaluation purposes, we have compared a standard four-layer CNN as well as a sophisticated FitNet4 architecture against their cross-modal variants on the CIFAR-10 and CIFAR-100 datasets with differing percentages of the training data being removed, and find that at lower levels of data availability, the X-CNNs significantly outperform their baselines (typically providing a 2-6% benefit, depending on the dataset size and whether data augmentation is used), while still maintaining an edge on all of the full dataset tests.
机译:在本文中,我们提出了交叉模式卷积神经网络(X-CNN),这是一种新型的生物启发型CNN架构,将梯度下降的专用CNN视为大型网络拓扑中的单个处理单元,同时允许不受约束的信息网络的类似隐藏层之间的流量和/或权重共享,从而概括了已经建立的神经网络集成概念(信息通常只能在各个网络的输出层之间流动)。各个组成网络经过单独设计,可以学习各自输入数据子集上的输出功能,然后在每次合并操作之后引入它们之间的交叉连接,以定期进行它们之间的信息交换。将知识注入模型(通过事先通过领域知识或无监督方法对输入数据进行划分)可望在稀疏数据环境中产生最大的回报,而稀疏数据环境通常不适合训练CNN。出于评估目的,我们将标准的四层CNN以及复杂的FitNet4架构与CIFAR-10和CIFAR-100数据集上的交叉模式变量进行了比较,并删除了不同百分比的训练数据,并发现较低的数据可用性水平,X-CNN明显优于其基线(通常提供2-6%的收益,具体取决于数据集大小和是否使用数据扩充),同时仍在所有完整数据集测试中保持优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号