首页> 外文会议>IEEE International Conference on Big Data Analytics >Named Entity Recognition Using BERT with Whole World Masking in Cybersecurity Domain
【24h】

Named Entity Recognition Using BERT with Whole World Masking in Cybersecurity Domain

机译:使用BERT与全世界屏蔽中的NERTITY识别在网络安全域中

获取原文

摘要

Nowadays the amount of cybersecurity data grows quickly on the Internet, however most of them are textual unstructured data, which is hard for security analysis to understand in time and is not suitable for automated security systems to directly use. The automated and real-time switching of cybersecurity information from unstructured text sources to structured representations can help the cyber threat intelligence analysis know the cyber situation better. Named Entity Recognition (NER) is able to convert unstructured data into structured data. Recently, a language representation model named Bidirectional Encoder Representations from Transformers (BERT) has achieved great improvements among different NLP tasks. In this paper, we apply BERT and its improved version BERT with whole world masking (BERTwwm) to the NER task for cybersecurity. We combine the BERT model with the BiLSTM-CRF architecture, and the experiment reveals that our method achieves greater performance on the precision, recall, and F1 score compared with the state-of-the-art model whether on the overall entity or single entity.
机译:如今,网络安全数据的数量在互联网上快速增长,但大多数是文本非结构化数据,这对于安全性分析很难及时了解,并且不适合直接使用的自动化安全系统。从非结构化文本源到结构化表示的自动和实时切换网络安全信息可以帮助网络威胁情报分析知道网络情况更好。命名实体识别(ner)能够将非结构化数据转换为结构化数据。最近,来自变换器(BERT)的名为Bidirectional编码器表示的语言表示模型已经实现了不同的NLP任务之间的巨大改进。在本文中,我们用全世界掩蔽涂抹BERT及其改进的版本伯特(BERT wwm )对于网络安全的人任务。我们将BERT模型与BILSTM-CRF架构相结合,实验表明,与最先进的模型相比,我们的方法在精度,召回和F1分数上实现了更高的性能,无论是整个实体还是单一实体。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号