首页> 外文会议>Annual CHI conference on human factors in computing systems >Crowdsourcing Performance Evaluations of User Interfaces
【24h】

Crowdsourcing Performance Evaluations of User Interfaces

机译:用户界面的众包性能评估

获取原文

摘要

Online labor markets, such as Amazon's Mechanical Turk (MTurk), provide an attractive platform for conducting human subjects experiments because the relative ease of recruitment, low cost, and a diverse pool of potential participants enable larger-scale experimentation and faster experimental revision cycle compared to lab-based settings. However, because the experimenter gives up the direct control over the participants' environments and behavior, concerns about the quality of the data collected in online settings are pervasive. In mis paper, we investigate the feasibility of conducting online performance evaluations of user interfaces with anonymous, unsupervised, paid participants recruited via MTurk. We implemented three performance experiments to re-evaluate three previously well-studied user interface designs. We conducted each experiment both in lab and online with participants recruited via MTurk. The analysis of our results did not yield any evidence of significant or substantial differences in the data collected in the two settings: All statistically significant differences detected in lab were also present on MTurk and the effect sizes were similar. In addition, there were no significant differences between the two settings in the raw task completion times, error rates, consistency, or the rates of utilization of the novel interaction mechanisms introduced in the experiments. These results suggest that MTurk may be a productive setting for conducting performance evaluations of user interfaces providing a complementary approach to existing methodologies.
机译:网上劳动力市场,比如亚马逊的Mechanical Turk(MTurk),为进行人类受试者实验提供了一个有吸引力的平台,因为相对容易招聘,成本低,潜在的参与者提供多样化的实现更大规模的试验和更快的实验修订周期比较基于实验室的设置。然而,因为实验者放弃了参与者的环境和行为的直接控制,约在网上设置收集的数据的质量问题普遍存在。在MIS文章中,我们进行调查的通过MTurk招募匿名,无监督,付费参与者的用户界面的在线绩效评估的可行性。我们实现了三个性能实验,以重新评估先前3充分研究用户界面设计。我们进行每个实验都在实验室和在线参与者通过MTurk招募。我们的研究结果的分析,没有产生的在两个设置收集的数据显著或重大分歧的证据:在实验室中检测到的所有统计显著差异也存在于MTurk和影响的大小相似。此外,还有两个设置之间的原始任务完成时间,错误率,一致性在实验中引入了新的互动机制的利用率无差异显著,或。这些结果表明,MTurk可以是用于导通的,提供一个互补的方法现有的方法的用户界面性能评估生产性设置。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号