首页> 外文会议>ACM/IEEE-CS Joint Conference on Digital Libraries >Evaluating the quality of educational answers in community question-answering
【24h】

Evaluating the quality of educational answers in community question-answering

机译:在社区问答中评估教育答案的质量

获取原文

摘要

Community Question-Answering (CQA), where questions and answers are generated by peers, has become a popular method of information seeking in online environments. While the content repositories created through CQA sites have been used widely to support general purpose tasks, using them as online digital libraries that support educational needs is an emerging practice. Horizontal CQA services, such as Yahoo! Answers, and vertical CQA services, such as Brainly, are aiming to help students improve their learning process by answering their educational questions. In these services, receiving high quality answer(s) to a question is a critical factor not only for user satisfaction, but also for supporting learning. However, the questions are not necessarily answered by experts, and the askers may not have enough knowledge and skill to evaluate the quality of the answers they receive. This could be problematic when students build their own knowledge base by applying inaccurate information or knowledge acquired from online sources. Using moderators could alleviate this problem. However, a moderator's evaluation of answer quality may be inconsistent because it is based on their subjective assessments. Employing human assessors may also be insufficient due to the large amount of content available on a CQA site. To address these issues, we propose a framework for automatically assessing the quality of answers. This is achieved by integrating different groups of features - personal, community-based, textual, and contextual - to build a classification model and determine what constitutes answer quality. To test this evaluation framework, we collected more than 10 million educational answers posted by more than 3 million users on Brainly's United States and Poland sites. The experiments conducted on these datasets show that the model using Random Forest (RF) achieves more than 83% accuracy in identifying high quality of answers. In addition, the findings indicate that personal and community-based features have more prediction power in assessing answer quality. Our approach also achieves high values on other key metrics such as F1-score and Area under ROC curve. The work reported here can be useful in many other contexts where providing automatic quality assessment in a digital repository of textual information is paramount.
机译:社区问答(CQA)是由同伴产生的问题和答案,已经成为在线环境中寻找信息的一种流行方法。尽管通过CQA网站创建的内容存储库已被广泛用于支持通用任务,但将它们用作支持教育需求的在线数字图书馆已成为一种新兴实践。横向CQA服务,例如Yahoo!答案和垂直的CQA服务(例如Brainly)旨在通过回答教育问题来帮助学生改善学习过程。在这些服务中,获得问题的高质量答案不仅是用户满意度的关键因素,也是支持学习的关键因素。但是,问题不一定由专家回答,询问者可能没有足够的知识和技能来评估他们收到的答案的质量。当学生通过应用不正确的信息或从在线资源获取的知识来建立自己的知识库时,这可能会出现问题。使用主持人可以缓解此问题。但是,主持人对答案质量的评估可能会不一致,因为它基于他们的主观评估。由于CQA网站上提供了大量内容,因此聘用评估人员可能也不够。为了解决这些问题,我们提出了一个自动评估答案质量的框架。这是通过集成不同的功能组(个人功能,基于社区的功能,文本功能和上下文功能)来建立分类模型并确定构成答案质量的要素而实现的。为了测试此评估框架,我们收集了超过300万用户在Brainly的美国和波兰网站上发布的1000万个教育答案。在这些数据集上进行的实验表明,使用随机森林(RF)的模型在识别高质量答案时可达到83%以上的准确性。此外,研究结果表明,基于个人和社区的功能在评估答案质量方面具有更大的预测能力。我们的方法还可以在其他关键指标(例如F1得分和ROC曲线下的面积)上获得高价值。在许多其他情况下,在文本信息的数字存储库中提供自动质量评估至关重要时,此处报告的工作可能会很有用。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号