首页> 外文学位 >Dimensionality of responses to a reading comprehension assessment and its implications to scoring test takers on their reading proficiency.
【24h】

Dimensionality of responses to a reading comprehension assessment and its implications to scoring test takers on their reading proficiency.

机译:阅读理解评估的回应维度及其对应试者阅读能力评分的含义。

获取原文
获取原文并翻译 | 示例

摘要

This dissertation investigated dimensionality of responses to questions in a passage-based reading comprehension assessment and its impacts on decisions that will be made about test takers' reading proficiency based on their test scores. Test takers' responses to the reading paper of Certificate in Advanced English (CAE) were analyzed using two quantitative approaches -- i.e., confirmatory factor analyses (CFA) and item response theory (IRT) analyses. The following three research questions were addressed in the study. 1. To what extent do items nested within passages exhibit multidimensionality? 2. How does the multidimensionality, if found, affect item parameter estimates when different measurement models are applied to the data? More specifically, how much and in what ways will item parameter estimates be biased when a unidimensional IRT model is applied to multidimensional data? 3. To what extent are person ability estimates influenced by whether multidimensionality is taken into account in a measurement model applied? More specifically, are the ability estimates comparable when estimated from two measurement models that does and does not take into account multidimensionality, respectively?;In order to answer the first research question, four CFA models --- i.e., one-factor, second-order factor, bi-factor and correlated-uniqueness models --- were specified to operationalize claims about the factor structure of test takers' responses and then tested on the basis of statistical fit as well as in terms of their substantive interpretability. The CFA results indicated that the one-factor model does not fit to the data and also that the bifactor model is the model that can best represent inter-item relationships on the CAE reading paper. These results imply that the item responses on the CAE reading test are multidimensional and that at least two sources of item inter-correlations --- reading proficiency and the specific reading passage on which items are based --- should be taken into account in order to properly model the item responses on the test.;The second and third research questions addressed the consequences of applying different measurement models that does or does not take into account multidimensionality observed in previous CFA analyses. To answer these research questions, two IRT models -- unidimensional and bifactor IRT models --- were applied to the data and the results were compared in terms of item parameter and person ability estimates. The results showed that an IRT model that takes into account conditional dependence among nested items within a reading passage, namely the bifactor model, provided different item parameter estimates from a model that assumes unidimensionality. These differences were particularly pronounced for the item discrimination parameter estimation. Furthermore, the use of different IRT models resulted in differences in scaling test takers' abilities. The finding that test takers were scored differently in different measurement models is particularly noteworthy, when it comes to the use of test scores for making decisions about individual test takers.;Based on the aforementioned results, this dissertation stresses that dimensionality of test takers' responses should be taken into account when choosing the most appropriate measurement model to be applied. The dissertation concludes with implications for using language assessment results in the real world -- making use of test scores for an intended purpose in both meaningful and equitable ways.
机译:本文研究了基于段落的阅读理解评估中对问题回答的维度及其对应试者根据他们的考试成绩对阅读能力做出决定的影响。使用两种定量方法分析了考生对高级英语证书(CAE)阅读纸的回答-确认因子分析(CFA)和项目响应理论(IRT)分析。研究中解决了以下三个研究问题。 1.嵌套在段落中的项目在多大程度上表现出多维性? 2.如果将多维度量模型应用于数据,那么多维性(如果找到)将如何影响项目参数估计?更具体地说,当将一维IRT模型应用于多维数据时,项目参数估计会以多少方式偏移? 3.在所应用的度量模型中,是否考虑了多维性,人员能力估计在多大程度上受到影响?更具体地说,当从分别考虑和不考虑多维的两个测量模型进行估计时,能力估计是否具有可比性?为了回答第一个研究问题,四个CFA模型(即单因素,第二因素)指定了阶乘因子,双因子模型和相关唯一性模型,以使关于应试者反应的因子结构的主张可操作,然后在统计拟合的基础上对其实质解释性进行检验。 CFA结果表明,单因素模型不适合数据,而且双因素模型是最能代表CAE阅读论文中项目间关系的模型。这些结果表明,CAE阅读测试中的项目回答是多维的,并且应该考虑项目互相关性的至少两个来源-阅读能力和项目所基于的特定阅读段落-以便正确地对测试中的项目响应进行建模。为了回答这些研究问题,将两个IRT模型(一维IRT模型和双因素IRT模型)应用于数据,并根据项目参数和人员能力估计对结果进行了比较。结果表明,IRT模型考虑了阅读段落中嵌套项目之间的条件依赖性,即双因素模型,它提供了与假定一维模型的不同项目参数估计。这些差异对于项目判别参数估计尤为明显。此外,使用不同的IRT模型会导致扩展考生能力的差异。当涉及到使用考试成绩来决定各个应试者的决策时,应试者在不同的测量模型中获得不同分数的发现尤其值得一提。基于上述结果,本论文强调应试者回答的维度选择最合适的测量模型时应考虑到这一点。论文的结论是在现实世界中使用语言评估结果的含义-以有意义和公平的方式利用考试成绩达到预期目的。

著录项

  • 作者

    So, Youngsoon.;

  • 作者单位

    University of California, Los Angeles.;

  • 授予单位 University of California, Los Angeles.;
  • 学科 Education Tests and Measurements.;Education English as a Second Language.
  • 学位 Ph.D.
  • 年度 2010
  • 页码 115 p.
  • 总页数 115
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号