首页> 外文会议>Conference on Empirical Methods in Natural Language Processing >HABERTOR: An Efficient and Effective Deep Hatespeech Detector
【24h】

HABERTOR: An Efficient and Effective Deep Hatespeech Detector

机译:Habertor:一种高效且有效的Deep HatePeech检测器

获取原文

摘要

We present our HABERTOR model for detecting hatespeech in large scale user-generated content. Inspired by the recent success of the BERT model, we propose several modifications to BERT to enhance the performance on the downstream hatespeech classification task. HABERTOR inherits BERT's architecture, but is different in four aspects: (ⅰ) it generates its own vocabularies and is pre-trained from the scratch using the largest scale hatespeech dataset; (ⅱ) it consists of Quaternion-based factorized components, resulting in a much smaller number of parameters, faster training and inferencing, as well as less memory usage; (ⅲ) it uses our proposed multi-source ensemble heads with a pooling layer for separate input sources, to further enhance its effectiveness; and (ⅳ) it uses a regularized adversarial training with our proposed finegrained and adaptive noise magnitude to enhance its robustness. Through experiments on the large-scale real-world hatespeech dataset with 1.4M annotated comments, we show that HABERTOR works better than 15 state-of-the-art hatespeech detection methods, including fine-tuning Language Models. In particular, comparing with BERT, our HABERTOR is 4~5 times faster in the training/inferencing phase, uses less than 1/3 of the memory, and has better performance, even though we pre-train it by using less than 1 % of the number of words. Our generalizability analysis shows that HABERTOR transfers well to other unseen hatespeech datasets and is a more efficient and effective alternative to BERT for the hatespeech classification.
机译:我们介绍了我们的Habertor模型,用于在大规模用户生成的内容中检测HatesPeech。灵感来自最近的BERT模型的成功,我们提出了几种修改,以增强下游讨厌静音分类任务的性能。 Habertor继承了BERT的架构,但在四个方面是不同的:(Ⅰ)它产生自己的词汇表,并使用最大秤宾馆数据集从划痕预训练; (Ⅱ)由四元素为基础的分子化组成,导致比较少量的参数,更快的培训和推理,以及更少的内存使用情况; (Ⅲ)它采用我们所提出的多源集合头,采用汇集层进行单独的输入来源,以进一步提高其有效性; (ⅳ)它使用了正规化的对抗性培训,我们提出了精细的和自适应噪声幅度,以提高其鲁棒性。通过对大型现实世界HATEPeech数据集的实验,具有1.4M的注释评论,我们表明Habertor工作优于15个最先进的HatesPeech检测方法,包括微调语言模型。特别是与伯特相比,我们的Habertor在训练/推理阶段速度快4〜5倍,使用少于1/3的内存,并且具有更好的性能,即使我们通过不到1%预先训练它单词数量。我们的普遍性分析表明,Habertor对其他看不见的HatesPeech数据集进行了良好,并且是一个更有效且有效的替代贝尔宾馆分类。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号