首页> 外文期刊>Journal of medical Internet research >Crowdsourcing a Normative Natural Language Dataset: A Comparison of Amazon Mechanical Turk and In-Lab Data Collection
【24h】

Crowdsourcing a Normative Natural Language Dataset: A Comparison of Amazon Mechanical Turk and In-Lab Data Collection

机译:众包标准化自然语言数据集:Amazon Mechanical Turk与实验室数据收集的比较

获取原文
       

摘要

Background: Crowdsourcing has become a valuable method for collecting medical research data. This approach, recruiting through open calls on the Web, is particularly useful for assembling large normative datasets. However, it is not known how natural language datasets collected over the Web differ from those collected under controlled laboratory conditions.Objective: To compare the natural language responses obtained from a crowdsourced sample of participants with responses collected in a conventional laboratory setting from participants recruited according to specific age and gender criteria.Methods: We collected natural language descriptions of 200 half-minute movie clips, from Amazon Mechanical Turk workers (crowdsourced) and 60 participants recruited from the community (lab-sourced). Crowdsourced participants responded to as many clips as they wanted and typed their responses, whereas lab-sourced participants gave spoken responses to 40 clips, and their responses were transcribed. The content of the responses was evaluated using a take-one-out procedure, which compared responses to other responses to the same clip and to other clips, with a comparison of the average number of shared words.Results: In contrast to the 13 months of recruiting that was required to collect normative data from 60 lab-sourced participants (with specific demographic characteristics), only 34 days were needed to collect normative data from 99 crowdsourced participants (contributing a median of 22 responses). The majority of crowdsourced workers were female, and the median age was 35 years, lower than the lab-sourced median of 62 years but similar to the median age of the US population. The responses contributed by the crowdsourced participants were longer on average, that is, 33 words compared to 28 words (P<.001), and they used a less varied vocabulary. However, there was strong similarity in the words used to describe a particular clip between the two datasets, as a cross-dataset count of shared words showed (P<.001). Within both datasets, responses contained substantial relevant content, with more words in common with responses to the same clip than to other clips (P<.001). There was evidence that responses from female and older crowdsourced participants had more shared words (P=.004 and .01 respectively), whereas younger participants had higher numbers of shared words in the lab-sourced population (P=.01).Conclusions: Crowdsourcing is an effective approach to quickly and economically collect a large reliable dataset of normative natural language responses.
机译:背景:众包已成为收集医学研究数据的一种有价值的方法。通过在Web上进行公开调用招募的这种方法对于组装大型规范数据集特别有用。但是,尚不清楚通过Web收集的自然语言数据集与在受控实验室条件下收集的自然语言数据集有何不同。目的:比较从众包参与者的样本中获得的自然语言反应与在常规实验室环境中根据方法:我们收集了来自Amazon Mechanical Turk工作者(众包)和从社区招募的60名参与者(来自实验室)的200个半分钟电影剪辑的自然语言描述。众包参与者对自己想要的片段进行了回应,并输入了答案,而实验室参与者则对40个片段进行了口头回答,并转录了他们的回答。回答的内容采用外带一遍程序进行评估,该过程将回答与相同剪辑和其他剪辑的其他响应进行比较,并比较共享词的平均数量。结果:与13个月相比从60名来自实验室的参与者(具有特定人口统计学特征)收集规范数据所需的招募中,仅需要34天就可以从99名众包参与者收集规范数据(贡献22位中值)。多数众包工人是女性,中位年龄为35岁,低于实验室采集的中位年龄62岁,但与美国人口的中位年龄相近。众包参与者的回答平均更长,即33个单词,而28个单词(P <.001),并且他们使用的词汇量也较少。但是,用于描述两个数据集之间的特定剪辑的词有很强的相似性,因为共享词的跨数据集计数显示为(P <.001)。在两个数据集中,响应都包含实质性的相关内容,对同一剪辑的响应比对其他剪辑的响应更多的单词(P <.001)。有证据表明,来自女性和年龄较大的众包参与者的回答具有更多的共有词(分别为P = .004和.01),而来自实验室的人群中年轻的参与者则具有更高的共有词数量(P = .01)。众包是一种快速,经济地收集规范自然语言反应的大型可靠数据集的有效方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号