首页> 外文会议>International Conference on Natural Language Processing >Well-Behaved Transformer for Chinese Medical NER
【24h】

Well-Behaved Transformer for Chinese Medical NER

机译:中国医疗设备用性能良好的变压器

获取原文

摘要

Medical named entity recognition (NER) is an important task of clinical natural language processing (NLP). It is a hot issue in intelligent medicine research. Recently, the proposed Lattice-LSTM model has demonstrated that incorporating information of words in character sequence into character-level Chinese NER has achieved new benchmark results on the Chinese datasets in multiple other fields. However, due to the lattice structure is dynamic and complex. These lattice-based models are difficult to fully use of the GPUs parallel computing, which limits its application. In this work, we propose a Well-Behaved Transformer (WB-Transformer) model for Chinese medical named entity recognition, using a high-performance encoding strategy to separately encode the character of Chinese electronic medical records (EMRs) and words corresponding to the character. This reduces the impact of word segmentation errors while obtaining the word boundary information, and makes full use information of characters and words for Chinese medical NER. Experimental on three Chinese medical entity recognition datasets show that our proposed model outperforms other methods.
机译:医学命名实体识别是临床自然语言处理的一项重要任务。这是智能医学研究中的一个热点问题。最近,提出的Lattice LSTM模型表明,将字符序列中的单词信息合并到字符级中文NER中,在其他多个领域的中文数据集上取得了新的基准结果。然而,由于晶格结构是动态的和复杂的。这些基于网格的模型难以充分利用GPU并行计算,限制了其应用。在这项工作中,我们提出了一个用于中文医疗命名实体识别的行为良好的转换器(WB Transformer)模型,使用一种高性能的编码策略来分别编码中文电子病历(EMR)的字符和对应于该字符的单词。这在获取词边界信息的同时,减少了分词错误的影响,充分利用了汉字和词的信息,为中医药研究提供了新的思路。在三个中国医学实体识别数据集上的实验表明,我们提出的模型优于其他方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号