首页> 外文会议>Proceedings of the 12th ACM conference on electronic commerce. >Who Moderates the Moderators? Crowdsourcing Abuse Detection in User-Generated Content
【24h】

Who Moderates the Moderators? Crowdsourcing Abuse Detection in User-Generated Content

机译:谁主持主持人?用户生成内容中的众包滥用检测

获取原文
获取原文并翻译 | 示例

摘要

A large fraction of user-generated content on the Web. such as posts or comments on popular online forums, consists of abuse or spam. Due to the volume of contributions on popular sites, a few trusted moderators cannot identify all such abusive content, so viewer ratings of contributions must be used for moderation. But not all viewers who rate content are trustworthy and accurate. What is a principled approach to assigning trust and aggregating user ratings, in order to accurately identify abusive content? In this paper, we introduce a framework to address the problem of moderating online content using crowdsourced ratings. Our framework encompasses users who are untrustworthy or inaccurate to an unknown extent that is, both the content and the raters are of unknown quality. With no knowledge whatsoever about the raters, it is impossible to do better than a random estimate. We present efficient algorithms to accurately detect abuse that only require knowledge about the identity of a single 'good' agent, who rates contributions accurately more than half the time. We prove that our algorithm can infer the quality of contributions with error that rapidly converges to zero as the number of observations increases; we also numerically demonstrate that the algorithm has very high accuracy for much fewer observations. Finally, we analyze the robustness of our algorithms to manipulation by adversarial or strategic raters, an important issue in moderating online content, and quantify how the performance of the algorithm degrades with the number of manipulating agents.
机译:用户在网络上生成的大部分内容。例如受欢迎的在线论坛上的帖子或评论,其中包括滥用或垃圾邮件。由于流行站点上的贡献数量众多,一些受信任的主持人无法识别所有此类辱骂性内容,因此必须使用观众对贡献的评分进行审核。但是,并非所有对内容进行评分的观众都是值得信赖和准确的。为了准确识别滥用内容,分配信任和汇总用户评级的原则方法是什么?在本文中,我们介绍了一个框架,以解决使用众包评分对在线内容进行审核的问题。我们的框架涵盖了不信任或不精确的用户,即内容和评估者的质量均未知。毫无关于评估者的知识,不可能做得比随机估计更好。我们提出了有效的算法来准确检测滥用行为,而这些滥用行为只需要了解单个“好”代理的身份即可,后者对贡献的评分准确度超过了一半。我们证明了我们的算法可以推断出贡献的质量,并且随着观察数量的增加,该质量可以迅速收敛到零。我们还通过数值方法证明了该算法具有非常高的准确度,而观测值却少得多。最后,我们分析了算法对对抗性或战略评估者进行操纵的鲁棒性,这是审核在线内容时的重要问题,并量化了算法性能如何随着操纵代理的数量而降低。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号