首页> 外文会议>International conference on computational linguistics >Does Higher Order LSTM Have Better Accuracy for Segmenting and Labeling Sequence Data?
【24h】

Does Higher Order LSTM Have Better Accuracy for Segmenting and Labeling Sequence Data?

机译:高阶LSTM在分割和标记序列数据方面是否具有更好的准确性?

获取原文

摘要

Existing neural models usually predict the tag of the current token independent of the neighboring tags. The popular LSTM-CRF model considers the tag dependencies between every two consecutive tags. However, it is hard for existing neural models to take longer distance dependencies of tags into consideration. The scalability is mainly limited by the complex model structures and the cost of dynamic programming during training. In our work, we first design a new model called "high order LSTM" to predict multiple tags for the current token which contains not only the current tag but also the previous several tags. We call the number of tags in one prediction as "order". Then we propose a new method called Multi-Order BiLSTM (MO-BiLSTM) which combines low order and high order LSTMs together. MO-BiLSTM keeps the scalability to high order models with a pruning technique. We evaluate MO-BiLSTM on all-phrase chunking and NER datasets. Experiment results show that MO-BiLSTM achieves the state-of-the-art result in chunking and highly competitive results in two NER datasets.
机译:现有的神经模型通常独立于相邻标签来预测当前令牌的标签。流行的LSTM-CRF模型考虑每两个连续标签之间的标签依赖关系。但是,现有的神经模型很难考虑标签的更长距离依赖性。可伸缩性主要受复杂的模型结构和训练过程中动态编程成本的限制。在我们的工作中,我们首先设计一个称为“高阶LSTM”的新模型,以预测当前令牌的多个标签,该令牌不仅包含当前标签,还包含先前的几个标签。我们将一个预测中的标签数称为“顺序”。然后,我们提出了一种称为多阶BiLSTM(MO-BiLSTM)的新方法,该方法将低阶LSTM和高阶LSTM组合在一起。 MO-BiLSTM通过修剪技术保持了对高阶模型的可伸缩性。我们在全短语组块和NER数据集上评估MO-BiLSTM。实验结果表明,MO-BiLSTM在两个NER数据集中以分块方式获得了最先进的结果,并获得了极具竞争力的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号