...
首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Factorized Recurrent Neural Architectures for Longer Range Dependence
【24h】

Factorized Recurrent Neural Architectures for Longer Range Dependence

机译:分解式递归神经体系结构可实现更长距离的依赖

获取原文
           

摘要

The ability to capture Long Range Dependence (LRD) in a stochastic process is of prime importance in the context of predictive models. A sequential model with a longer-term memory is better able contextualize recent observations. In this article, we apply the theory of LRD stochastic processes to modern recurrent architectures, such as LSTMs and GRUs, and prove they do not provide LRD under assumptions sufficient for gradients to vanish. Motivated by an information-theoretic analysis, we provide a modified recurrent neural architecture that mitigates the issue of faulty memory through redundancy while keeping the compute time constant. Experimental results on a synthetic copy task, the Youtube-8m video classification task and a recommender system show that we enable better memorization and longer-term memory.
机译:在随机过程中捕获远程依赖关系(LRD)的能力在预测模型的背景下至关重要。具有长期记忆的顺序模型可以更好地将最近的观察结果与背景联系起来。在本文中,我们将LRD随机过程的理论应用于现代递归体系结构(例如LSTM和GRU),并证明它们在足以消除梯度的假设下不提供LRD。受信息理论分析的启发,我们提供了一种改进的循环神经体系结构,该体系结构通过冗余减轻了错误内存的问题,同时保持了计算时间不变。在合成复制任务,Youtube-8m视频分类任务和推荐系统上的实验结果表明,我们可以实现更好的记忆和长期记忆。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号