首页> 外文期刊>Expert systems with applications >Combating hate speech using an adaptive ensemble learning model with a case study on COVID-19
【24h】

Combating hate speech using an adaptive ensemble learning model with a case study on COVID-19

机译:使用自适应集合学习模型对抗仇恨言论,案例研究Covid-19

获取原文
获取原文并翻译 | 示例
           

摘要

Social media platforms generate an enormous amount of data every day. Millions of users engage themselves with the posts circulated on these platforms. Despite the social regulations and protocols imposed by these platforms, it is difficult to restrict some objectionable posts carrying hateful content. Automatic hate speech detection on social media platforms is an essential task that has not been solved efficiently despite multiple attempts by various researchers. It is a challenging task that involves identifying hateful content from social media posts. These posts may reveal hate outrageously, or they may be subjective to the user or a community. Relying on manual inspection delays the process, and the hateful content may remain available online for a long time. The current state-of-the-art methods for tackling hate speech perform well when tested on the same dataset but fail miserably on cross-datasets. Therefore, we propose an ensemble learning-based adaptive model for automatic hate speech detection, improving the cross-dataset generalization. The proposed expert model for hate speech detection works towards overcoming the strong user-bias present in the available annotated datasets. We conduct our experiments under various experimental setups and demonstrate the proposed model's efficacy on the latest issues such as COVID-19 and US presidential elections. In particular, the loss in performance observed under cross-dataset evaluation is the least among all the models. Also, while restricting the maximum number of tweets per user, we incur no drop in performance.
机译:社交媒体平台每天都会产生大量数据。数百万用户使用在这些平台上分发的帖子。尽管这些平台施加了社会法规和协议,但很难限制一些患有可恶内容的令人反感的职位。社交媒体平台上的自动仇恨语音检测是尽管各种研究人员尝试多次尝试,但尚未有效地解决的重要任务。这是一个具有挑战性的任务,涉及从社交媒体帖子中识别仇恨内容。这些帖子可能令人兴奋地揭示仇恨,或者它们可能是用户或社区的主观。依靠手动检查延迟过程,仇恨内容可能会长期可用。当在同一数据集上测试时,当前用于解决仇恨语音的最先进方法,但在交叉数据集中失败。因此,我们提出了一种基于集合的自动讨厌语音检测的自适应自适应模型,提高了交叉数据集泛化。讨论语音检测的建议专家模型用于克服可用注释数据集中存在的强大用户偏见。我们在各种实验设置下进行实验,并展示拟议的模型对Covid-19和美国总统选举等最新问题的效果。特别是,在交叉数据集评估下观察到的性能损失是所有模型中的至少。此外,在限制每个用户的最大推文的最大数量时,我们不会在性能下降。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号