Research Article

Examining the impact of violations of local item independence assumption on test equating methods

Volume: 12 Number: 3 September 4, 2025
EN

Examining the impact of violations of local item independence assumption on test equating methods

Abstract

This study investigates the impact of violating the local item independence assumption by loading certain items onto a second dimension on test equating errors in unidimensional and dichotomous tests. The research was designed as a simulation study, using data generated based on the PISA 2018 mathematics exam. Analyses were conducted under 36 different conditions, varying by sample sizes (250, 1000, and 5000), test lengths (20, 40, and 60 items), and proportions of items loaded onto the second dimension (0%, 15%, 30%, and 50%). A "random groups design" was used, resulting in the creation of 3600 datasets through 100 replications. The results revealed that the equating methods based on classical test theory (CTT) showed varying levels of error depending on the error types and conditions. Among the item response theory (IRT) scale transformation methods, the Stocking-Lord method produced the least error values and was the least affected by violations of the local independence assumption. Additionally, the observed score equating method demonstrated lower root mean square error (RMSE) values than the true score equating method and was less affected by local independence violations. The SS-MIRT observed score equating method yielded lower RMSE values compared to the other methods and was found to be more robust against the violation of the local independence assumption.

Keywords

References

  1. Aiken, L.R. (2000). Psychological testing and assesment (10th ed.). Allyn and Bacon.
  2. Aksekioğlu, B. (2017). Madde tepki kuramına dayalı test eşitleme yöntemlerinin karşılaştırılması: PISA 2012 fen testi örneği [Comparison of test equating methods based on item response theory: PISA 2012 science test sample]. [Master's Thesis, Akdeniz University]. Higher Education Institution National Thesis Center.
  3. Akour, M.M.M. (2006). A comparison of various equipercentile and kernel equating methods under the random groups design. [Doctoral Dissertation, Graduate College of The University of Iowa]. ProQuest Dissertations and Theses Global.
  4. Albano, A.D. (2016). equate: An R package for observed-score linking and equating. Journal of Statistical Software, 74, 1-36. https://doi.org/10.18637/jss.v074.i08
  5. Angoff, W.H. (1984). Scales, norms, and equivalent scores. Educational Testing Service.
  6. Aşiret, S. (2014). Küçük Örneklemlerde test eşitleme yöntemlerinin çeşitli faktörlere göre incelenmesi [Factors affecting the test equating method using small samples]. [Master's Thesis, Mersin University]. Higher Education Institution National Thesis Center.
  7. Atar, B., & Yeşiltaş, G. (2017) Çok boyutlu eşitleme yöntemlerinin eşdeğer olmayan gruplarda ortak madde deseni için performanslarının incelenmesi [Investigation of the performance of multidimensional equating procedures for common-item nonequivalent groups design]. Journal of Measurement and Evaluation in Education and Psychology, 8(4), 421-434. https://doi.org/10.21031/epod.335284
  8. Baker, F.B., & Al‐Karni, A. (1991). A comparison of two procedures for computing IRT equating coefficients. Journal of Educational Measurement, 28(2), 147-162.

Details

Primary Language

English

Subjects

Measurement Theories and Applications in Education and Psychology , Similation Study

Journal Section

Research Article

Early Pub Date

July 21, 2025

Publication Date

September 4, 2025

Submission Date

October 7, 2024

Acceptance Date

February 21, 2025

Published in Issue

Year 2025 Volume: 12 Number: 3

APA
Doğuyurt, M. F., & Tan, Ş. (2025). Examining the impact of violations of local item independence assumption on test equating methods. International Journal of Assessment Tools in Education, 12(3), 629-661. https://doi.org/10.21449/ijate.1562627

23823             23825             23824