首页> 外文会议>International Conference on Progress in Informatics and Computing >Lexicon-based semi-CRF for Chinese clinical text word segmentation
【24h】

Lexicon-based semi-CRF for Chinese clinical text word segmentation

机译:基于词汇的SEMI-CRF用于中国临床文本词分割

获取原文

摘要

Word segmentation is in most cases a base for text analysis and absolutely vital to the accuracy of subsequent natural language processing (NLP) tasks. While word segmentation for normal text has been intensively studied and quite a few algorithms have been proposed, these algorithms however do not work well in special fields, e.g., in clinical text analysis. Besides, most state-of-the-art methods have difficulties in identifying out-of-vocabulary (OOV) words. For these two reasons, in this paper, we propose a semi-supervised CRF (semi-CRF) algorithm for Chinese clinical text word segmentation. Semi-CRF is implemented by modifying the learning objective so as to adapt for partial labeled data. Training data are obtained by applying a bidirectional lexicon matching scheme. A modified Viterbi algorithm using lexicon matching scheme is also proposed for word segmentation on raw sentences. Experiments show that our model has a precision of 93.88% on test data and outperforms two popular open source Chinese word segmentation tools i.e., HanLP and THULAC. By using lexicon, our model is able to be adapted for other domain text word segmentation.
机译:在大多数情况下,单词分割是文本分析的基础,对后续自然语言处理(NLP)任务的准确性绝对至关重要。虽然已经深入研究了正常文本的单词分割,并且已经提出了相当多的算法,但这些算法在特殊领域,例如,在临床文本分析中不起作用。此外,大多数最先进的方法在识别词汇外(OOV)单词方面具有困难。由于这两个原因,在本文中,我们提出了一种半监督CRF(半CRF)算法,用于中国临床文本词分割。通过修改学习目标来实现半CRF,以便适应部分标记的数据。通过应用双向词典匹配方案获得训练数据。还提出了一种修改的Viterbi算法,用于原始句子上的单词分段。实验表明,我们的模型在测试数据上具有93.88 %的精确度,并且优于两个流行的开源中文字分割工具即,Hanlp和Thulac。通过使用Lexicon,我们的模型能够适用于其他域文本字分段。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号