首页> 外文会议>ACM/IEEE-CS Joint Conference on Digital Libraries >Evaluating the quality of educational answers in community question-answering
【24h】

Evaluating the quality of educational answers in community question-answering

机译:评估社区质询回答中的教育答案质量

获取原文

摘要

Community Question-Answering (CQA), where questions and answers are generated by peers, has become a popular method of information seeking in online environments. While the content repositories created through CQA sites have been used widely to support general purpose tasks, using them as online digital libraries that support educational needs is an emerging practice. Horizontal CQA services, such as Yahoo! Answers, and vertical CQA services, such as Brainly, are aiming to help students improve their learning process by answering their educational questions. In these services, receiving high quality answer(s) to a question is a critical factor not only for user satisfaction, but also for supporting learning. However, the questions are not necessarily answered by experts, and the askers may not have enough knowledge and skill to evaluate the quality of the answers they receive. This could be problematic when students build their own knowledge base by applying inaccurate information or knowledge acquired from online sources. Using moderators could alleviate this problem. However, a moderator's evaluation of answer quality may be inconsistent because it is based on their subjective assessments. Employing human assessors may also be insufficient due to the large amount of content available on a CQA site. To address these issues, we propose a framework for automatically assessing the quality of answers. This is achieved by integrating different groups of features - personal, community-based, textual, and contextual - to build a classification model and determine what constitutes answer quality. To test this evaluation framework, we collected more than 10 million educational answers posted by more than 3 million users on Brainly's United States and Poland sites. The experiments conducted on these datasets show that the model using Random Forest (RF) achieves more than 83% accuracy in identifying high quality of answers. In addition, the findings indicate that personal and community-based features have more prediction power in assessing answer quality. Our approach also achieves high values on other key metrics such as F1-score and Area under ROC curve. The work reported here can be useful in many other contexts where providing automatic quality assessment in a digital repository of textual information is paramount.
机译:社区问答(CQA),在对等方生成的问题和答案,已成为在线环境中寻求的热门信息方法。虽然通过CQA站点创建的内容存储库已被广泛用于支持通用任务,但在线数字图书馆支持支持教育需求是新兴的惯例。水平CQA服务,如雅虎!答案和垂直的CQA服务,例如大脑,旨在通过回答他们的教育问题来帮助学生改善他们的学习过程。在这些服务中,接受了一个问题的高质量答案是不仅为用户满意度的关键因素,而且是为了支持学习。但是,专家不一定回答问题,提问者可能没有足够的知识和技能来评估他们收到的答案的质量。当学生通过应用从在线来源获得的不准确的信息或知识构建自己的知识库时,这可能是有问题的。使用主持人可以缓解此问题。然而,主持人对答案质量的评估可能是不一致的,因为它基于其主观评估。由于CQA网站上可用的大量含量,雇用人类评估员也可能不足。要解决这些问题,我们提出了一个框架,用于自动评估答案的质量。这是通过集成不同的特征组 - 个人,社区,文本和语境 - 构建分类模型并确定构成答案质量的组。为了测试这一评估框架,我们收集了超过1000万的教育答案,在大脑的美国和波兰地点上有超过300万用户发布。在这些数据集上进行的实验表明,使用随机森林(RF)的模型在识别高质量的答案时达到了83%的准确性。此外,该研究结果表明,基于个人和社区的特征在评估答案质量方面具有更多的预测力。我们的方法还在其他关键指标上实现了高价值,例如ROC曲线下的F1分数和面积。这里报告的工作在许多其他背景下都有用,其中在文本信息的数字存储库中提供自动质量评估是至关重要的。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号