首页> 外文期刊>Wireless communications & mobile computing >A Multichannel Biomedical Named Entity Recognition Model Based on Multitask Learning and Contextualized Word Representations
【24h】

A Multichannel Biomedical Named Entity Recognition Model Based on Multitask Learning and Contextualized Word Representations

机译:基于多任务学习和上下文化字表示的多通道生物医学名为实体识别模型

获取原文
           

摘要

As the biomedical literature increases exponentially, biomedical named entity recognition (BNER) has become an important task in biomedical information extraction. In the previous studies based on deep learning, pretrained word embedding becomes an indispensable part of the neural network models, effectively improving their performance. However, the biomedical literature typically contains numerous polysemous and ambiguous words. Using fixed pretrained word representations is not appropriate. Therefore, this paper adopts the pretrained embeddings from language models (ELMo) to generate dynamic word embeddings according to context. In addition, in order to avoid the problem of insufficient training data in specific fields and introduce richer input representations, we propose a multitask learning multichannel bidirectional gated recurrent unit (BiGRU) model. Multiple feature representations (e.g., word-level, contextualized word-level, character-level) are, respectively, or collectively fed into the different channels. Manual participation and feature engineering can be avoided through automatic capturing features in BiGRU. In merge layer, multiple methods are designed to integrate the outputs of multichannel BiGRU. We combine BiGRU with the conditional random field (CRF) to address labels’ dependence in sequence labeling. Moreover, we introduce the auxiliary corpora with same entity types for the main corpora to be evaluated in multitask learning framework, then train our model on these separate corpora and share parameters with each other. Our model obtains promising results on the JNLPBA and NCBI-disease corpora, with F1-scores of 76.0% and 88.7%, respectively. The latter achieves the best performance among reported existing feature-based models.
机译:随着生物医学文献呈指数增长,生物医学命名实体识别(BNER)已成为生物医学信息提取的重要任务。在以前的基于深度学习的研究中,预先灌注的单词嵌入成为神经网络模型不可或缺的一部分,有效提高了它们的性能。然而,生物医学文献通常包含许多多殖民和含糊不清的词。使用固定的预磨词表示不合适。因此,本文采用语言模型(ELMO)的预染色嵌入,根据上下文生成动态字嵌入。另外,为了避免特定领域中训练数据不足的问题并引入更丰富的输入表示,我们提出了一种多址学习多通道双向门控复发单元(BigRU)模型。多个特征表示(例如,单词级,语境化词级,字符级别)分别或集体馈入不同的信道。通过在Bigru中的自动捕获功能,可以避免手动参与和特征工程。在合并层中,旨在集成多声道Bigru的输出。我们将Bigru与条件随机字段(CRF)组合以解决标签在序列标记中的依赖性。此外,我们介绍了在多任务学习框架中进行评估的主语料的同一实体类型的辅助Corpora,然后在这些单独的语料库上培训我们的模型,并互相共享参数。我们的模型在JNLPBA和NCBI-Discult Corpora上获得了有希望的结果,分别具有76.0%和88.7%的F1分。后者达到了报告的基于功能的模型中的最佳性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号