...
首页> 外文期刊>Information Processing & Management >How well do hate speech, toxicity, abusive and offensive language classification models generalize across datasets?
【24h】

How well do hate speech, toxicity, abusive and offensive language classification models generalize across datasets?

机译:仇恨言语,毒性,滥用和令人反感的语言分类模型如何概括到数据集?

获取原文
获取原文并翻译 | 示例
           

摘要

A considerable body of research deals with the automatic identification of hate speech and related phenomena. However, cross-dataset model generalization remains a challenge. In this context, we address two still open central questions: (ⅰ) to what extent does the generalization depend on the model and the composition and annotation of the training data in terms of different categories?, and (ⅱ) do specific features of the datasets or models influence the generalization potential? To answer (ⅰ), we experiment with BERT, ALBERT, fastText, and SVM models trained on nine common public English datasets, whose class (or category) labels are standardized (and thus made comparable), in intra- and cross-dataset setups. The experiments show that indeed the generalization varies from model to model and that some of the categories (e.g., 'toxic', 'abusive', or 'offensive') serve better as cross-dataset training categories than others (e.g., 'hate speech'). To answer (ⅱ), we use a Random Forest model for assessing the relevance of different model and dataset features during the prediction of the performance of 450 BERT, 450 ALBERT, 450 fastText, and 348 SVM binary abusive language classifiers (1698 in total). We find that in order to generalize well, a model already needs to perform well in an intra-dataset scenario. Furthermore, we find that some other parameters are equally decisive for the success of the generalization, including, e.g., the training and target categories and the percentage of the out-of-domain vocabulary.
机译:相当大的研究涉及仇恨言论的自动识别和相关现象。但是,跨数据集模型概括仍然是一个挑战。在这种情况下,我们解决了两个仍然开放的中央问题:(Ⅰ)泛化在多大程度上取决于模型和组成以及在不同类别方面的培训数据的构图和注释(Ⅱ)做出具体特征数据集或模型影响泛化潜力?答案(Ⅰ),我们在九个常见公共英语数据集上进行培训,艾伯特,FastText和SVM模型,其类(或类别)标签在内部和交叉数据集设置中标准化(并因此进行了可比) 。实验表明,概述从模型变化到模型,其中一些类别(例如,“毒性”,“辱骂”或“攻击性”)优于交叉数据集培训类别(例如,'仇恨讲话')。要回答(Ⅱ),我们使用随机森林模型来评估不同模型和数据集特征在预测450 BERT,450 Albert,450 FastText和348个SVM二进制滥用语言分类器的情况下的相关模型和数据集特征的相关性(总共1698年) 。我们发现,为了概括,模型已经需要在一个数据集内场景中表现良好。此外,我们发现一些其他参数对于概括的成功同样决定性,包括例如培训和目标类别以及域外词汇表的百分比。

著录项

  • 来源
    《Information Processing & Management》 |2021年第3期|102524.1-102524.17|共17页
  • 作者单位

    Natural Language Processing Group Department of Communication and Information Technologies Pompeu Fabra University Spain;

    Natural Language Processing Group Department of Communication and Information Technologies Pompeu Fabra University Spain;

    Catalan Institute for Research and Advanced Studies (ICREA) Spain Natural Language Processing Group Department of Communication and Information Technologies Pompeu Fabra University Spain;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Hate speech; Offensive language; Classification; Generalization;

    机译:仇恨言论;令人反感的语言;分类;概括;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号