Reliability of the Analytic Rubric and Checklist for the Assessment of Story Writing Skills: G and Decision Study in Generalizability Theory
Abstract
The purpose of study is to examine the reliability of analytical rubrics and checklists developed for the assessment of story writing skills by means of generalizability theory. The study group consisted of 52 students attending the 5th grade at primary school and 20 raters in Mersin University. The G study was carried out with the fully crossed hxpxg (story x rater x performance task) design, where the scoring keys were determined as fix facet. Decision Study was carried out by changing the task facet conditions. As a result, it was observed in both scoring keys that the sources of variance related to the stories had a high variance percentage in the main effects while "hp (story and rater interaction effects)" a high variance percentage in the interaction effects. The highest variance in the design belongs to the interaction effect "hpg (story, rater and performance task interaction effects)". This can be an indicator for the existence of different sources of variability and error, which are not included in the design. Examining the G and phi coefficients calculated for both scoring keys, it was determined that scoring with analytic rubrics is more reliable and generalizable. According to the decision studies, it was decided that the number of tasks used in this study is to be most appropriate.
Keywords
Kaynakça
- Aktas, M. (2013). An Investigation of the Reliability of the Scores Obtained Through Rating the Same Performance Task with Three Different Techniques by Different Numbers of Raters According to Generalizability Theory (Unpublished master’s thesis). Mersin University/Institute of Education Sciences, Mersin, Turkey.
- Atilgan, H. (2004). A Research on The Comparability of Generalizability Theory and Multivariate Rasch Model (Unpublished doctorate thesis). Hacettepe University/ Institute of Social Sciences, Ankara, Turkey.
- Bachman, L. F., Lynch, B. K. & Mason, M. (1995). Investigating Variability in Asks and Rater Judgements in a Performance Test of Foreign Language Speaking. Language Testing, 12, 238-257.
- Branthwaite, A., Trueman, M., & Berrisford, T. (1981). Unreliability of Marking: Further Evidence and a Possible Explanation. Education Review, 33(1), 41-46.
- Breland, H. M. (1983). The Direct Assessment of Writing Skill: A Measurement Review, College Board Report No. 83-6, ETS RR No. 83-32, New York: College Examination Board.
- Brennan, R. L. (1992). Elements of Generalizability Theory (rev. ed.). Iowa City IA: ACT.
- Brennan, R. L. (2001). Generalizability Theory. New York: Springer-Verlag.
- Brookhart, S. M. (1999). The Art and Science of Classroom Assessment: the Missing Part of Pedagogy. Ashe-Eric Higher Education Report (Vol. 27, No.1). Washington, DC: The George Washington University, Graduate School of Education and Human Development.
Ayrıntılar
Birincil Dil
İngilizce
Konular
Eğitim Üzerine Çalışmalar
Bölüm
Araştırma Makalesi
Yayımlanma Tarihi
15 Ocak 2019
Gönderilme Tarihi
30 Ekim 2018
Kabul Tarihi
18 Aralık 2018
Yayımlandığı Sayı
Yıl 1970 Cilt: 8 Sayı: 1