首页> 外文会议>International conference on neural information processing >Bi-directional LSTM Recurrent Neural Network for Chinese Word Segmentation
【24h】

Bi-directional LSTM Recurrent Neural Network for Chinese Word Segmentation

机译:双向LSTM递归神经网络用于中文分词

获取原文

摘要

Recurrent neural network (RNN) has been broadly applied to natural language process (NLP) problems. This kind of neural network is designed for modeling sequential data and has been testified to be quite efficient in sequential tagging tasks. In this paper, we propose to use bi-directional RNN with long short-term memory (LSTM) units for Chinese word segmentation, which is a crucial task for modeling Chinese sentences and articles. Classical methods focus on designing and combining hand-craft features from context, whereas bi-directional LSTM network (BLSTM) does not need any prior knowledge or pre-designing, and is expert in creating hierarchical feature representation of contextual information from both directions. Experiment result shows that our approach gets state-of-the-art performance in word segmentation on both traditional Chinese datasets and simplified Chinese datasets.
机译:递归神经网络(RNN)已广泛应用于自然语言过程(NLP)问题。这种神经网络是为建模顺序数据而设计的,并已被证明在顺序标记任务中非常有效。在本文中,我们建议使用带有长短期记忆(LSTM)单元的双向RNN进行中文分词,这是建模中文句子和文章的关键任务。经典方法着重于根据上下文设计和组合手工特征,而双向LSTM网络(BLSTM)不需要任何先验知识或预先设计,并且擅长从两个方向创建上下文信息的分层特征表示。实验结果表明,我们的方法在传统中文数据集和简体中文数据集上均具有最新的分词性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号