期刊文献+

结合题目作答时间的计算机化自适应测验选题方法 被引量:3

The Use of Response Time in Item Selection of Computerized Adaptive Testing
下载PDF
导出
摘要 计算机形式的测验能够记录考生在测验中的题目作答时间(RT),作为一种重要的辅助信息来源,RT对于测验开发和管理具有重要的价值,特别是在计算机化自适应测验(CAT)领域。本文简要介绍了RT在CAT选题方面应用并作以简评,分析了这些技术在实践中的可行性。最后,探讨了当前RT应用于CAT选题存在的问题以及可以进一步开展的研究方向。 The computer-based test enables the examinee’s response time(RT)to be recorded accurately.As an important source of auxiliary information,RT have an important potential value for test development and management,especially in the field of Computerized Adaptive Testing(CAT).With the collection of RT,the CAT assessment process can be further improved in terms of precision,fairness,and minimizing costs.It is widely known that item selection is the key step of CAT,which reflects its"adaptive"characteristics.The traditional CAT item selection algorithm does not consider RT information,this is unfavorable for test management and may lead to biased assessment results.This paper synthetically and briefly introduces the application of RT in the item selection of CAT and analyzes the feasibility of these techniques in practice,which makes the readers have a specific and clear understanding of the potential value of RT in CAT.Since item selection in CAT is based on the candidate's ability estimation(except for the selection of initial items),the improvement of ability estimation can also be considered as an indirect improvement of the item selection.Therefore,this paper divides relevant methods into two categories:(1)indirect improvement of item selection by RTs(ability estimation)and(2)direct improvement of item selection by RT(item selection method).Generally,a majority of tests are a mixture of speed and power components,while the RT provide information not only examinees’ability but also item characteristics.In the past decades,a lot of models for RT and response accuracy(RA)has been proposed(e.g.,Thissen,1983;Wang&Hanson,2005;van der Linden,2007),which makes it possible to use RT to improve the accuracy of ability estimation in CAT,and the item selection is further improved(van der Linden,2008).In general,examinees with the same ability level may need different time to complete an item(van der Linden et al.,1999),and the response time of an examinee for different items may also be different because some items are usually more time consuming than others(Veldkamp,2016).Test speededness results in examinees taking different amounts of time to complete a test.However,most standardized tests often set a specific time for practical administration purposes,when candidates are pressured by the time limit,they may improve the response speed at the expense of accuracy(Entink et al.,2009),which leads to biased ability estimation.Therefore,it is necessary to eliminate the influence of the speed factor for the test whose main goal is to evaluate ability,and this is more in line with the unidimensional hypothesis of IRT.However,the conventional item selection methods didn’t take this into account,and RT information should be introduced into the process of item selection to address this problem(van der Linden et al.,1999;Fan et al.,2012).With the development of measurement theory and technology,researchers hope to get richer diagnostic information about an examinee from the test,rather than simply evaluating him on an abstract scale,and the application of RTs is a promissing approach.
作者 郭治辰 汪大勋 蔡艳 涂冬波 Guo Zhichen;Wang Daxun;Cai Yan;Tu Dongbo(School of Psychology,Jiangxi Normal University,Nanchang,330022)
出处 《心理科学》 CSSCI CSCD 北大核心 2021年第5期1241-1248,共8页 Journal of Psychological Science
基金 国家自然科学基金(31960186、31760288、31660278)项目的资助。
关键词 计算机化自适应测验 题目作答时间 能力估计 选题方法 题目曝光 测验时间 computerized adaptive testing item response time ability estimation item exposure item selection method test time
  • 相关文献

参考文献4

二级参考文献27

  • 1Azzalini, A. (1985). A class of distributions which includes the normal ones. Scandinavian Journal of Statisges, 12, 171-178.
  • 2Baker, F. B. & Kim, S. H. (2004). Item response theory: Parameter estimation techniques. New York: Marcel Dekker.
  • 3Kang, T., & Cohen, A. S. (2007). IRT model selection methods for dichotomous items. Applied Psychological Measurement, 31, 331-358.
  • 4Klein Entink, R. H., Fox, J. P., & van der Linden, W. J. (2009). A multivariate multilevel approach to the modeling of accuracy and speed of test takers. Psyehomettlka, 74, 21 --48.
  • 5Klein Entink, R. H., van der Linden, W. J., & Fox, J. P. (2009). A Box-Cox normal model for response times. British Journal of Mathematical and Statistical Psychology, 62, 621-640.
  • 6Lee, Y. H., & Chen, H. (2011). A review of recent response-time analyses in educational testing. Psychological Test and Assessment Modeling, 53, 359- 379.
  • 7Li, F. M., Cohen, A. S., Kim, S. H., & Cho, S. J. (2009). Model selection methods for mixture, dichotomous IRT models. Applied Psychological Measurement, 33, 353-373.
  • 8I.oeys, T., Rosseel, Y., & Batena, K. (2011). A joint modeling approach for reaction time and accuracy in psycholinguistic experiments. Psyehometffka, 76, 487- 503.
  • 9Meyer, J. P. (2010). A mixture Rasch model with item response time components. Applied Psycbological Measurement, 34, 521-538.
  • 10Meng, X. B., Tan, J., & Chang, H. H. (2015). A conditional joint modeling approach for locally dependent item responses and response times. Journal of Educational Measurement, 52, 1-27.

共引文献19

同被引文献24

引证文献3

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部