首页> 外文期刊>International journal on digital libraries >Assessing the quality of answers autonomously in community question-answering
【24h】

Assessing the quality of answers autonomously in community question-answering

机译:在社区问答中自主评估答案的质量

获取原文
获取原文并翻译 | 示例
       

摘要

Community question-answering (CQA) has become a popular method of online information seeking. Within these services, peers ask questions and create answers to those questions. For some time, content repositories created through CQA sites have widely supported general-purpose tasks; however, they can also be used as online digital libraries that satisfy specific needs related to education. Horizontal CQA services, such as Yahoo! Answers, and vertical CQA services, such as Brainly, aim to help students improve their learning process via Q&A exchanges. In addition, Stack Overflow—another vertical CQA—serves a similar purpose but specifically focuses on topics relevant to programmers. Receiving high-quality answer(s) to a posed CQA query is a critical factor to both user satisfaction and supported learning in these services. This process can be impeded when experts do not answer questions and/or askers do not have the knowledge and skills needed to evaluate the quality of the answers they receive. Such circumstances may cause learners to construct a faulty knowledge base by applying inaccurate information acquired from online sources. Though site moderators could alleviate this problem by surveying answer quality, their subjective assessments may cause evaluations to be inconsistent. Another potential solution lies in human assessors, though they may also be insufficient due to the large amount of content available on a CQA site. The following study addresses these issues by proposing a framework for automatically assessing answer quality. We accomplish this by integrating different groups of features—personal, community-based, textual, and contextual—to build a classification model and determine what constitutes answer quality. We collected more than 10 million educational answers posted by more than 3 million users on Brainly and 7.7 million answers on Stack Overflow to test this evaluation framework. The experiments conducted on these data sets show that the model using random forest achieves high accuracy in identifying high-quality answers. Findings also indicate that personal and community-based features have more prediction power in assessing answer quality. Additionally, other key metrics such as F1-score and area under ROC curve achieve high values with our approach. The work reported here can be useful in many other contexts that strive to provide automatic quality assessment in a digital repository.
机译:社区问答(CQA)已成为在线信息搜索的一种流行方法。在这些服务中,同伴会提出问题并为这些问题创建答案。一段时间以来,通过CQA网站创建的内容存储库已广泛支持通用任务。但是,它们也可以用作满足与教育相关的特定需求的在线数字图书馆。横向CQA服务,例如Yahoo!答案和垂直的CQA服务(例如Brainly)旨在通过问答交流来帮助学生改善学习过程。另外,堆栈溢出(另一种垂直CQA)具有类似的目的,但专门针对与程序员相关的主题。接收对提出的CQA查询的高质量答案是这些服务中用户满意度和支持的学习的关键因素。当专家不回答问题和/或提问者不具备评估他们收到的答案的质量所需的知识和技能时,可能会阻止此过程。这种情况可能会导致学习者通过应用从在线资源中获取的不正确信息来构建错误的知识库。尽管站点主持人可以通过调查答案质量来缓解此问题,但他们的主观评估可能会导致评估结果不一致。另一个潜在的解决方案在于人工评估者,尽管由于CQA网站上可用的大量内容,评估者可能也不够。以下研究通过提出一个自动评估答案质量的框架来解决这些问题。我们通过集成不同组的功能(个人功能,基于社区的功能,文本功能和上下文功能)来实现此目的,以建立分类模型并确定构成答案质量的要素。我们收集了超过300万用户在Brainly上发布的1000万个教育答案,以及Stack Overflow上的770万个答案,以测试此评估框架。在这些数据集上进行的实验表明,使用随机森林的模型在识别高质量答案方面具有很高的准确性。研究结果还表明,基于个人和社区的功能在评估答案质量方面具有更大的预测能力。此外,其他关键指标(例如F1得分和ROC曲线下的面积)也可以通过我们的方法获得较高的价值。此处报告的工作在许多其他试图在数字存储库中提供自动质量评估的环境中可能很有用。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号