首页> 外文会议>International Workshop on Semantic Evaluation >IIITG-ADBU at SemEval-2020 Task 12: Comparison of BERT and BiLSTM in Detecting Offensive Language
【24h】

IIITG-ADBU at SemEval-2020 Task 12: Comparison of BERT and BiLSTM in Detecting Offensive Language

机译:iiitg-adbu在Semeval-2020任务12:伯特和Bilstm在检测攻击性语言中的比较

获取原文

摘要

Task 12 of SemEval 2020 consisted of 3 subtasks, namely offensive language identification (Subtask A), categorization of offense type (Subtask B), and offense target identification (Subtask C). This paper presents the results our classifiers obtained for the English language in the 3 subtasks. The classifiers used by us were BERT and BiLSTM. On the test set, our BERT classifier obtained a macro F1 score of 0.90707 for subtask A, and 0.65279 for subtask B. The BiLSTM classifier obtained a macro F1 score of 0.57565 for subtask C. The paper also performs an analysis of the errors made by our classifiers. We conjecture that the presence of a few misleading instances in the dataset is affecting the performance of the classifiers. Our analysis also discusses the need for temporal context and world knowledge to determine the offensiveness of a few comments.
机译:Semeval 2020的任务12由3个子任务组成,即令人反感的语言识别(子任务A),攻击类型(SubTask B)的分类和冒犯目标识别(子任务C)。 本文介绍了我们在3个子任务中获取英语语言的分类器的结果。 我们使用的分类器是BERT和BILSTM。 在测试集上,我们的BERT分类器获得了SubTask A的0.9070分的宏F1得分,以及用于子任务B的0.65279。Bilstm分类器获得0.57565的宏F1分数为0.57565,用于子任务C.本文还执行对误差的分析 我们的分类器。 我们猜测数据集中的一些误导性实例的存在是影响分类器的性能。 我们的分析还讨论了需要时间背景和世界知识,以确定几个评论的攻击性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号