首页> 外文期刊>Neural computing & applications >A deep neural network-based model for named entity recognition for Hindi language
【24h】

A deep neural network-based model for named entity recognition for Hindi language

机译:基于深度神经网络的印地语语言名称实体识别模型

获取原文
获取原文并翻译 | 示例
       

摘要

The aim of this work is to develop efficient named entity recognition from the given text that in turn improves the performance of the systems that use natural language processing (NLP). The performance of IoT-based devices such as Alexa and Cortana significantly depends upon an efficient NLP model. To increase the capability of the smart IoT devices in comprehending the natural language, named entity recognition (NER) tools play an important role in these devices. In general, the NER is a two-step process that initially the proper nouns are identified from text and then classify them into predefined categories of entities such as person, location, measure, organization and time. NER is often performed as a subtask while processing natural languages which increases the accuracy level of a NLP task. In this paper, we propose deep neural network architecture for named entity recognition for the resource-scarce language Hindi, based on convolutional neural network (CNN), bidirectional long short-term memory (Bi-LSTM) neural network and conditional random field (CRF). In the proposed approach, initially, we use skip-gram word2vec model and GloVe model to represent words in semantic vectors which are further used in different deep neural network-based architectures. In the proposed approach, we use character- and word-level embedding to represent the text that includes information at fine-grained level. Due to the use of character-level embeddings, the proposed model is robust for the out-of-vocabulary words. Experimental results show that the combination of Bi-LSTM, CNN and CRF algorithms performs better as compared to the other baseline methods such as recurrent neural network, long short-term memory and Bi-LSTM individually.
机译:这项工作的目的是从给定的文本开发有效的命名实体识别,反过来改善了使用自然语言处理(NLP)的系统的性能。基于IOT的设备如Alexa和Cortana的性能显着取决于高效的NLP模型。为了提高智能物联网设备在理解自然语言时,命名实体识别(ner)工具在这些设备中发挥着重要作用。通常,ner是一个两步的过程,最初是从文本中识别的正确名词,然后将它们分为预定义的类别,例如人,位置,测量,组织和时间。 NER通常作为子任务进行,同时处理自然语言,这增加了NLP任务的精度级别。在本文中,我们提出了深度神经网络架构,用于基于卷积神经网络(CNN),双向长短期记忆(Bi-LSTM)神经网络和条件随机场(CRF )。在拟议的方法中,首先,我们使用Skip-gram Word2Vec模型和手套模型来表示语义向量中的单词,这些载体中进一步用于基于深度神经网络的架构。在所提出的方法中,我们使用字符和字级嵌入来表示包含细粒度级别信息的文本。由于使用字符级嵌入式,所提出的模型对于词汇外单词是强大的。实验结果表明,与其他基线方法(如经常性神经网络,长短短期记忆和双LSTM)相比,Bi-LSTM,CNN和CRF算法的组合更好地执行更好。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号