【24h】

Tensor Switching Networks

机译:张量交换网络

获取原文

摘要

We present a novel neural network algorithm, the Tensor Switching (TS) network, which generalizes the Rectified Linear Unit (ReLU) nonlinearity to tensor-valued hidden units. The TS network copies its entire input vector to different locations in an expanded representation, with the location determined by its hidden unit activity. In this way, even a simple linear readout from the TS representation can implement a highly expressive deep-network-like function. The TS network hence avoids the vanishing gradient problem by construction, at the cost of larger representation size. We develop several methods to train the TS network, including equivalent kernels for infinitely wide and deep TS networks, a one-pass linear learning algorithm, and two backpropagation-inspired representation learning algorithms. Our experimental results demonstrate that the TS network is indeed more expressive and consistently learns faster than standard ReLU networks.
机译:我们提出了一种新颖的神经网络算法,即张量交换(TS)网络,该算法将整流线性单位(ReLU)非线性泛化为张量值隐藏单位。 TS网络将其整个输入矢量复制到一个扩展的表示形式中的不同位置,该位置由其隐藏的单位活动确定。这样,即使从TS表示中进行简单的线性读出也可以实现高度表达的类深层网络功能。因此,TS网络避免了构造上消失的梯度问题,以较大的表示尺寸为代价。我们开发了几种训练TS网络的方法,包括用于无限宽和深TS网络的等效内核,一遍线性学习算法和两种反向传播启发式表示学习算法。我们的实验结果表明,与标准ReLU网络相比,TS网络确实更具表达力,并且学习速度始终如一。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号