...
首页> 外文期刊>Computer speech and language >A Framework for pre-training hidden-unit conditional random fields and its extension to long short term memory networks
【24h】

A Framework for pre-training hidden-unit conditional random fields and its extension to long short term memory networks

机译:预训练隐藏单元条件随机字段的框架及其对长期短期记忆网络的扩展

获取原文
获取原文并翻译 | 示例
           

摘要

In this paper, we introduce a simple unsupervised framework for pre-training hidden-unit conditional random fields (HUCRFs), i.e., learning initial parameter estimates for HUCRFs prior to supervised training.Our framework exploits the model structure of HUCRFs to make effective use of unlabeled data from the same domain or labeled data from a different domain. The key idea is to use the separation of HUCRF parameters between observations and labels: this allows us to pre-train observation parameters independently of label parameters. Pre-training is achieved by creating pseudo-labels from such resources. In the case of unlabeled data, we cluster observations and use the resulting clusters as pseudo-labels. Observation parameters can be trained on these resources and then transferred to initialize the supervised training process on the target labeled data. Experiments on various sequence labeling tasks demonstrate that the proposed pre-training method consistently yields significant improvement in performance. The core idea could be extended to other learning techniques including deep learning. We applied the proposed technique to recurrent neural networks (RNN) with long short term memory (LSTM) architecture and obtained similar gains.
机译:在本文中,我们介绍了一个简单的无监督框架,用于预训练隐藏单元条件随机字段(HUCRF),即在有监督训练之前学习HUCRF的初始参数估计。我们的框架利用HUCRF的模型结构来有效利用来自相同域的未标记数据或来自不同域的被标记数据。关键思想是在观察值和标签之间使用HUCRF参数的分离:这使我们能够独立于标签参数来预训练观察参数。通过从此类资源创建伪标签来实现预训练。在未标记数据的情况下,我们对观察结果进行聚类,并将所得聚类用作伪标签。可以在这些资源上训练观察参数,然后将其传输以初始化对目标标记数据的监督训练过程。在各种序列标记任务上的实验表明,所提出的预训练方法始终可以显着提高性能。核心思想可以扩展到其他学习技术,包括深度学习。我们将提出的技术应用于具有长期短期记忆(LSTM)架构的递归神经网络(RNN),并获得了相似的收益。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号