首页> 外文会议>Annual CHI conference on human factors in computing systems >Crowdsourcing Performance Evaluations of User Interfaces
【24h】

Crowdsourcing Performance Evaluations of User Interfaces

机译:用户界面的众包性能评估

获取原文

摘要

Online labor markets, such as Amazon's Mechanical Turk (MTurk), provide an attractive platform for conducting human subjects experiments because the relative ease of recruitment, low cost, and a diverse pool of potential participants enable larger-scale experimentation and faster experimental revision cycle compared to lab-based settings. However, because the experimenter gives up the direct control over the participants' environments and behavior, concerns about the quality of the data collected in online settings are pervasive. In mis paper, we investigate the feasibility of conducting online performance evaluations of user interfaces with anonymous, unsupervised, paid participants recruited via MTurk. We implemented three performance experiments to re-evaluate three previously well-studied user interface designs. We conducted each experiment both in lab and online with participants recruited via MTurk. The analysis of our results did not yield any evidence of significant or substantial differences in the data collected in the two settings: All statistically significant differences detected in lab were also present on MTurk and the effect sizes were similar. In addition, there were no significant differences between the two settings in the raw task completion times, error rates, consistency, or the rates of utilization of the novel interaction mechanisms introduced in the experiments. These results suggest that MTurk may be a productive setting for conducting performance evaluations of user interfaces providing a complementary approach to existing methodologies.
机译:像Amazon的Mechanical Turk(MTurk)这样的在线劳动力市场为进行人体实验提供了一个有吸引力的平台,因为与之相比,招聘相对容易,成本低廉,并且潜在参与者人数众多,因此可以进行大规模实验并加快实验修订周期到基于实验室的设置。但是,由于实验者放弃了对参与者的环境和行为的直接控制,因此对在线设置中收集的数据质量的担忧无处不在。在错误的文件中,我们调查了通过MTurk招募的匿名,无监督,付费参与者对用户界面进行在线性能评估的可行性。我们实施了三个性能实验,以重新评估三个以前经过充分研究的用户界面设计。我们在实验室和在线进行了每个实验,并通过MTurk招募了参与者。对我们的结果进行的分析没有得出任何证据表明在两种情况下收集到的数据存在显着或实质性差异:在实验室中检测到的所有统计学上显着差异也出现在MTurk上,并且效果大小相似。此外,在原始任务完成时间,错误率,一致性或实验中引入的新型交互机制的利用率方面,两种设置之间没有显着差异。这些结果表明,MTurk可能是进行用户界面性能评估的有效设置,为现有方法提供了一种补充方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号