Bu araştırmanın amacı üst düzey düşünme becerilerini ölçebilecek çoktan seçmeli madde hazırlarken dikkat edilmesi gereken noktaları göstermek ve var olan genel yanılgının aksine çoktan seçmeli maddelerle üst düşünme becerilerinin de ölçülebileceğine dikkat çekmektir. Sınıf içi değerlendirmelerden geniş ölçekli testlere kadar eğitim alanının birçok kısmında büyük bir yer kapladığı tartışılmaz bir gerçek olan çoktan seçmeli maddeleri daha üst düzey becerilere hitap edecek şekilde yapılandırmak oldukça önemlidir. Ancak bu tür maddelerin yaygın kullanımına karşın
kaliteli ve üst düzey becerileri ölçebilecek çoktan seçmeli maddeler oluşturmak hâlâ öğretmenler, eğitimciler ve test geliştiriciler için zor olmaya devam etmektedir. Çalışmada üst düzey becerileri
ölçebilecek çoktan seçmeli madde yazımı için literatürde önerilen tekniklere yer verilmiş ve bir dizi pratik öneriler sunulmuştur. Bahsedilen tekniklerle çoktan seçmeli maddelerin daha üst düzey
düşünme becerilerini ölçebileceği sonucuna varılmıştır. Bu tekniklerin tanıtılması ve derlenmesiyle oluşturulan bu çalışmanın hem öğrencilere hem de test geliştiricilere yarar sağlayabileceği
düşünülmektedir.
1. Kelly FJ. The Kansas silent reading tests. J Educ Psychol. 1916;7(2):63-
80. [CrossRef]
2. Lukhele R, Thissen D, Wainer H. On the relative value of multiplechoice, constructed response, and examinee-selected items
on two achievement tests. J Educ Meas. 1994;31(3):234-250.
[CrossRef]
3. Scully D. Constructing multiple-choice items to measure higherorder thinking. Pract Assess Res Eval. 2017;22(4):1-13.
4. Burton SJ, Sudweeks RR, Merrill PF, Wood B. How to Prepare Better
Multiple-Choice Test Items: Guidelines for University Faculty. Utah:
Brigham Young University Testing Services and the Department of
Instructional Science; 1991;1-33.
5. Lemons PP, Lemons JD. Questions for assessing higher-order
cognitive skills: it’s not just Bloom’s. Sci Educ. 2013;12(1):47-58.
[CrossRef]
6. Istiyono E. The development of physics essay test for higher order
thinking skills in junior high school. Proceeding of International Conference on Research, Implementation and Education of Mathematics and Sciences 2014 May 18-20; Yogyakarta State University.
7. Newmann FM. Higher order thinking in the teaching of social studies:
Connections between theory and practice. In: Voss J, Perkins D,
Segal J, eds. Informal Reasoning and Education. Hillsdale, NJ: Lawrence Erlbaum; 1991:381-400.
8. Daly WM. Critical thinking as an outcome of nursing education. What
is it? Why is it important to nursing practice? J Adv Nurs.
1998;28(2):323-331. [CrossRef]
9. Schroeder JM. A Study of Improving Critical Thinking Skills with Multiple Choice Tests and First Semester Associate Degree Nursing Students. Doctoral dissertation. Minneapolis, MN: Capella University;
2007.
10. Dickinson M. Writing multiple-choice questions for higher-level
thinking. Learning Solutions Magazine. Available at: https://campbel
lms.typepad.com/files/writing-multiple-choice-questions-for-highe
r-level-thinking-by-mike-dickinson-_-learning-solutions-magazine.
pdf. Accessed March 2, 2020.
11. Forehand M. (2005) Bloom’s Taxonomy: Original and Revised. In M.
Orey (Ed.), Emerging Perspectives on Learning, Teaching, and Technology (E-Book). https://textbookequity.org/Textbooks/Orey_Emergin_
Perspectives_Learning.pdf http://epltt.coe.uga.edu/index.php?title
=Bloom%27s_Taxonomy. Accessed January 20, 2011.
12. Bloom B, Englehart M, Furst E, Hill W, Krathwohl D. A Taxonomy of
Educational Objectives, Handbook I: Cognitive Domain. New York:
David McKay Company; 1956.
13. Bloom B. Reflections on the development and use of the taxonomy.
In: In A. L, Sosniak L, eds. Bloom’s Taxonomy: A Forty-Year Retrospective. Chicago: The University of Chicago Press; 1994:1-8.
14. Anderson LW, Krathwohl DR, Airasian PW, et al. A Taxonomy for
Learning, Teaching, and Assessing: A Revision of Bloom’s Taxonomy
of Educational Objectives. complete ed. New York: Longman; 2001.
15. Wilson LO. Anderson and Krathwohl–Bloom’s taxonomy revised.
Understanding the new version of Bloom’s taxonomy. Available at:
https://quincycollege.edu/content/uploads/Anderson-and-Krath
wohl_Revised-Blooms-Taxonomy.pdf. Accessed March 11, 2020.
16. Wiggins G. Five unfortunate misunderstanding that almost all educators have about Bloom’s Taxonomy [blog post]. Available at: https
://grantwiggins.wordpress.com/2015/03/04/5-unfortunate-misun
derstandingsthat-almost-all-educators-have-about-blooms-taxo
nomy/. Accessed March 11, 2020.
17. Jovanovska J. Designing effective multiple-choice questions for
assessing learning outcomes. Infotheca, J Digit Humanit. 2018;18(1):
25-42.
18. Morrison S, Free KW. Writing multiple-choice test items that promote and measure critical thinking. J Nurs Educ. 2001;40(1):17-24.
[CrossRef]
19. Crowe A, Dirks C, Wenderoth MP. Biology in bloom: Implementing
Bloom’s taxonomy to enhance student learning in biology. CBE—
Life. CBE Life Sci Educ. 2008;7(4):368-381. [CrossRef].
20. Brookhart SM. How to Assess Higher-Order Thinking Skills in Your
Classroom. Alexandria: ASCD; 2010.
21. Clay B. A Short Guide to Writing Effective Test Questions. Lawrence:
Kansas Curriculum Center, University of Kansas; 2001.
22. Azer SA. Assessment in a problem-based learning course: Twelve
tips for constructing multiple choice questions that test students’
cognitive skills. Biochem Mol Biol Educ. 2003;31(6):428-434.
[CrossRef]
23. Burns ER. “Anatomizing” reversed: Use of examination questions
that foster use of higher order learning skills by students. Anat Sci
Educ. 2010;3(6):330-334. [CrossRef]
24. Haladyna TM, Downing SM, Rodriguez MC. A review of multiplechoice item-writing guidelines for classroom assessment. Appl Meas
Educ. 2002;15(3):309-333. [CrossRef]
25. Collins J. Education techniques for lifelong learning: Writing
multiple-choice questions for continuing medical education activities and self-assessment modules. RadioGraphics. 2006;26(2):
543-551. [CrossRef]
26. Zaidi NB, Hwang C, Scott S, Stallard S, Purkiss J, Hortsch M. Climbing
Bloom’s taxonomy pyramid: Lessons from a graduate histology
course. Anat Sci Educ. 2017;10(5):456-464. [CrossRef]
27. Abdalla ME, Gaffar AM, Suliman RA. Constructing A-type Multiple
Choice Questions (MCQs): Step by Step Manual. Blueprints in Health
Profession Education Series. 2011. Retrieved from https://www.res
earchgate.net/publication/235914088
28. Haladyna TM. Developing and Validating MuItiple-Choice Test Items.
Hillsdale, NJ: Lawrence Erlbaum Associates; 1994.
29. Ascalon ME, Meyers LS, Davis BW, Smits N. Distractor similarity and
item-stem structure: effects on item difficulty. Appl Meas Educ.
2007;20(2):153-170. [CrossRef]
30. DiBattista D, Kurzawa L. Examination of the quality of multiplechoice items on classroom tests. cjsotl-rcacea. 2011;2(2):1-23.
[CrossRef]
31. Hancock GR. Cognitive complexity and the comparability of multiple-choice and constructed-response test formats. J Exp Educ.
1994;62(2):143-157. [CrossRef]
32. Ericsson KA, Simon H. Protocol Analysis: Verbal Reports as Data. rev.
ed. Cambridge, MA: MIT Press; 1993.
33. Çelik H, Kocabıyık E, Sönmezer Ü. Kocaeli İl Milli Eğitim Müdürlüğü
Ölçme ve Değerlendirme Merkezi tarafından hazırlanan 6. ve 7. Sınıf
matematik dersi ortak yazılı sınav sorularının madde türlerine ve
yenilenmiş Bloom taksonomisine göre değerlendirilmesi. Ege Bilimsel Araştırmalar Derg. 2020;3(1):28-53.
34. Gökulu A. Fen ve teknoloji öğretmenlerinin yazılı sınav soruları ile
TEOG sınavlarında sorulan fen ve teknoloji sorularının yenilenmiş
Bloom taksonomisine göre incelenmesi. Route Educational and
Social Science Journal. 2015;2(2):434-446.
35. Koç H, Sönmez ÖF, Çiftçi T. YGS ve LYS sınavlarındaki coğrafya
sorularının Bloom taksonomisi bilişsel alan düzeyi açısından analizi.
J Black Sea Stud. 2013;9(36):257-275.
36. Özmen H. ÖSS sınavlarındaki kimya sorularının konu alanlarına ve
Bloom taksonomisine göre incelenmesi. Eurasian J Educ Res (EJER).
1990-2005;21:187-199.
37. Badat T, Usgu G, Dinler E, Bayramlar K, Yakut Y. Çoktan seçmeli
sınavlarda kullanılan ölçme ve değerlendirme sisteminin uygulanması:
Madde analiz örneği. Hacettepe Univ Fac Health Sci J. 2020;7(3):
285-295.
38. Cansever Z, Acemoğlu H, Avşar Ü, Hoşoğlu S. Tıp fakültesindeki çoktan seçmeli sınav sorularının değerlendirilmesi. Tıp Eğitimi Dünyası.
2016;14(44):44-55. [CrossRef]
39. Khorshid L, Zaybak A, Güneş ÜY, et al. Hemşirelik yüksekokulu çoktan
seçmeli sınav sorularının analizi. Ege Univ Hemşirelik Fak Derg.
2010;26(2):37-48.
40. Mahjabeen W, Alam S, Hassan U, et al. Difficulty index, discrimination
index and distractor efficiency in multiple choice questions. Annals
PIMS-Shaheed Zulfiqar Ali Bhutto Medical University. 2018;13(4):
310-315.
41. Özcan KV, Aydoğan Y, Bulut İ. Gaziosmanpaşa Üniversitesi Tıp Fakültesinde uygulanan çoktan seçmeli sınavların betimsel analizi.
Gaziosmanpaşa Univ Tıp Fak Derg. 2014;6(4):281-294.
42. Topal K, Aybek H, Kara CO, Büke A, Aybek Z. PAÜ Tıp Fakültesi dönem
ı öğrencilerine 2006-2007 eğitim ve öğretim yılında uygulanan çoktan seçmeli sınavların madde ve test analizleri. Pamukkale Tıp Derg.
2008;1(3):120-126.
43. Tiemeier AM, Stacy ZA, Burke JM. Using multiple choice questions
written at various Bloom’s taxonomy levels to evaluate student performance across a therapeutics sequence. Innov Pharm. 2011;2(2):1-
11. [CrossRef]
44. Simkin MG, Kuechler WL. Multiple-choice tests and student understanding: What is the connection? Decis Sci J Innov Educ.
2005;3(1):73-98. [CrossRe
Suggestions for Designing Multiple Choice Items for High-Order Thinking Skills with Examples from the Healthcare Field
Year 2022,
Volume: 25 Issue: 2, 117 - 125, 30.06.2022
The purpose of this study is to show points that should be considered when preparing multiplechoice items that can measure high-level thinking skills and to draw attention to the fact that the higher thinking skills can be measured with multiple-choice items. From classroom evaluations to large-scale tests, it is very important to configure multiple-choice items which is an indisputable fact that many parts of the training area occupy a big place, to address higher-level skills. Nevertheless, despite the widespread use of such items, it is still difficult for teachers, educators, and test developers to write multiple-choice items that can measure quality and high-level skills. In the study, the techniques proposed in the literature for writing multiple-choice items that can measure high-level skills are given and a series of practical suggestions are presented. It has been concluded that the multiple-choice items can measure higher level thinking skills with the mentioned techniques. It is thought that this study, which is formed by introducing and compiling these techniques, can benefit both students and test developers in various contexts.
1. Kelly FJ. The Kansas silent reading tests. J Educ Psychol. 1916;7(2):63-
80. [CrossRef]
2. Lukhele R, Thissen D, Wainer H. On the relative value of multiplechoice, constructed response, and examinee-selected items
on two achievement tests. J Educ Meas. 1994;31(3):234-250.
[CrossRef]
3. Scully D. Constructing multiple-choice items to measure higherorder thinking. Pract Assess Res Eval. 2017;22(4):1-13.
4. Burton SJ, Sudweeks RR, Merrill PF, Wood B. How to Prepare Better
Multiple-Choice Test Items: Guidelines for University Faculty. Utah:
Brigham Young University Testing Services and the Department of
Instructional Science; 1991;1-33.
5. Lemons PP, Lemons JD. Questions for assessing higher-order
cognitive skills: it’s not just Bloom’s. Sci Educ. 2013;12(1):47-58.
[CrossRef]
6. Istiyono E. The development of physics essay test for higher order
thinking skills in junior high school. Proceeding of International Conference on Research, Implementation and Education of Mathematics and Sciences 2014 May 18-20; Yogyakarta State University.
7. Newmann FM. Higher order thinking in the teaching of social studies:
Connections between theory and practice. In: Voss J, Perkins D,
Segal J, eds. Informal Reasoning and Education. Hillsdale, NJ: Lawrence Erlbaum; 1991:381-400.
8. Daly WM. Critical thinking as an outcome of nursing education. What
is it? Why is it important to nursing practice? J Adv Nurs.
1998;28(2):323-331. [CrossRef]
9. Schroeder JM. A Study of Improving Critical Thinking Skills with Multiple Choice Tests and First Semester Associate Degree Nursing Students. Doctoral dissertation. Minneapolis, MN: Capella University;
2007.
10. Dickinson M. Writing multiple-choice questions for higher-level
thinking. Learning Solutions Magazine. Available at: https://campbel
lms.typepad.com/files/writing-multiple-choice-questions-for-highe
r-level-thinking-by-mike-dickinson-_-learning-solutions-magazine.
pdf. Accessed March 2, 2020.
11. Forehand M. (2005) Bloom’s Taxonomy: Original and Revised. In M.
Orey (Ed.), Emerging Perspectives on Learning, Teaching, and Technology (E-Book). https://textbookequity.org/Textbooks/Orey_Emergin_
Perspectives_Learning.pdf http://epltt.coe.uga.edu/index.php?title
=Bloom%27s_Taxonomy. Accessed January 20, 2011.
12. Bloom B, Englehart M, Furst E, Hill W, Krathwohl D. A Taxonomy of
Educational Objectives, Handbook I: Cognitive Domain. New York:
David McKay Company; 1956.
13. Bloom B. Reflections on the development and use of the taxonomy.
In: In A. L, Sosniak L, eds. Bloom’s Taxonomy: A Forty-Year Retrospective. Chicago: The University of Chicago Press; 1994:1-8.
14. Anderson LW, Krathwohl DR, Airasian PW, et al. A Taxonomy for
Learning, Teaching, and Assessing: A Revision of Bloom’s Taxonomy
of Educational Objectives. complete ed. New York: Longman; 2001.
15. Wilson LO. Anderson and Krathwohl–Bloom’s taxonomy revised.
Understanding the new version of Bloom’s taxonomy. Available at:
https://quincycollege.edu/content/uploads/Anderson-and-Krath
wohl_Revised-Blooms-Taxonomy.pdf. Accessed March 11, 2020.
16. Wiggins G. Five unfortunate misunderstanding that almost all educators have about Bloom’s Taxonomy [blog post]. Available at: https
://grantwiggins.wordpress.com/2015/03/04/5-unfortunate-misun
derstandingsthat-almost-all-educators-have-about-blooms-taxo
nomy/. Accessed March 11, 2020.
17. Jovanovska J. Designing effective multiple-choice questions for
assessing learning outcomes. Infotheca, J Digit Humanit. 2018;18(1):
25-42.
18. Morrison S, Free KW. Writing multiple-choice test items that promote and measure critical thinking. J Nurs Educ. 2001;40(1):17-24.
[CrossRef]
19. Crowe A, Dirks C, Wenderoth MP. Biology in bloom: Implementing
Bloom’s taxonomy to enhance student learning in biology. CBE—
Life. CBE Life Sci Educ. 2008;7(4):368-381. [CrossRef].
20. Brookhart SM. How to Assess Higher-Order Thinking Skills in Your
Classroom. Alexandria: ASCD; 2010.
21. Clay B. A Short Guide to Writing Effective Test Questions. Lawrence:
Kansas Curriculum Center, University of Kansas; 2001.
22. Azer SA. Assessment in a problem-based learning course: Twelve
tips for constructing multiple choice questions that test students’
cognitive skills. Biochem Mol Biol Educ. 2003;31(6):428-434.
[CrossRef]
23. Burns ER. “Anatomizing” reversed: Use of examination questions
that foster use of higher order learning skills by students. Anat Sci
Educ. 2010;3(6):330-334. [CrossRef]
24. Haladyna TM, Downing SM, Rodriguez MC. A review of multiplechoice item-writing guidelines for classroom assessment. Appl Meas
Educ. 2002;15(3):309-333. [CrossRef]
25. Collins J. Education techniques for lifelong learning: Writing
multiple-choice questions for continuing medical education activities and self-assessment modules. RadioGraphics. 2006;26(2):
543-551. [CrossRef]
26. Zaidi NB, Hwang C, Scott S, Stallard S, Purkiss J, Hortsch M. Climbing
Bloom’s taxonomy pyramid: Lessons from a graduate histology
course. Anat Sci Educ. 2017;10(5):456-464. [CrossRef]
27. Abdalla ME, Gaffar AM, Suliman RA. Constructing A-type Multiple
Choice Questions (MCQs): Step by Step Manual. Blueprints in Health
Profession Education Series. 2011. Retrieved from https://www.res
earchgate.net/publication/235914088
28. Haladyna TM. Developing and Validating MuItiple-Choice Test Items.
Hillsdale, NJ: Lawrence Erlbaum Associates; 1994.
29. Ascalon ME, Meyers LS, Davis BW, Smits N. Distractor similarity and
item-stem structure: effects on item difficulty. Appl Meas Educ.
2007;20(2):153-170. [CrossRef]
30. DiBattista D, Kurzawa L. Examination of the quality of multiplechoice items on classroom tests. cjsotl-rcacea. 2011;2(2):1-23.
[CrossRef]
31. Hancock GR. Cognitive complexity and the comparability of multiple-choice and constructed-response test formats. J Exp Educ.
1994;62(2):143-157. [CrossRef]
32. Ericsson KA, Simon H. Protocol Analysis: Verbal Reports as Data. rev.
ed. Cambridge, MA: MIT Press; 1993.
33. Çelik H, Kocabıyık E, Sönmezer Ü. Kocaeli İl Milli Eğitim Müdürlüğü
Ölçme ve Değerlendirme Merkezi tarafından hazırlanan 6. ve 7. Sınıf
matematik dersi ortak yazılı sınav sorularının madde türlerine ve
yenilenmiş Bloom taksonomisine göre değerlendirilmesi. Ege Bilimsel Araştırmalar Derg. 2020;3(1):28-53.
34. Gökulu A. Fen ve teknoloji öğretmenlerinin yazılı sınav soruları ile
TEOG sınavlarında sorulan fen ve teknoloji sorularının yenilenmiş
Bloom taksonomisine göre incelenmesi. Route Educational and
Social Science Journal. 2015;2(2):434-446.
35. Koç H, Sönmez ÖF, Çiftçi T. YGS ve LYS sınavlarındaki coğrafya
sorularının Bloom taksonomisi bilişsel alan düzeyi açısından analizi.
J Black Sea Stud. 2013;9(36):257-275.
36. Özmen H. ÖSS sınavlarındaki kimya sorularının konu alanlarına ve
Bloom taksonomisine göre incelenmesi. Eurasian J Educ Res (EJER).
1990-2005;21:187-199.
37. Badat T, Usgu G, Dinler E, Bayramlar K, Yakut Y. Çoktan seçmeli
sınavlarda kullanılan ölçme ve değerlendirme sisteminin uygulanması:
Madde analiz örneği. Hacettepe Univ Fac Health Sci J. 2020;7(3):
285-295.
38. Cansever Z, Acemoğlu H, Avşar Ü, Hoşoğlu S. Tıp fakültesindeki çoktan seçmeli sınav sorularının değerlendirilmesi. Tıp Eğitimi Dünyası.
2016;14(44):44-55. [CrossRef]
39. Khorshid L, Zaybak A, Güneş ÜY, et al. Hemşirelik yüksekokulu çoktan
seçmeli sınav sorularının analizi. Ege Univ Hemşirelik Fak Derg.
2010;26(2):37-48.
40. Mahjabeen W, Alam S, Hassan U, et al. Difficulty index, discrimination
index and distractor efficiency in multiple choice questions. Annals
PIMS-Shaheed Zulfiqar Ali Bhutto Medical University. 2018;13(4):
310-315.
41. Özcan KV, Aydoğan Y, Bulut İ. Gaziosmanpaşa Üniversitesi Tıp Fakültesinde uygulanan çoktan seçmeli sınavların betimsel analizi.
Gaziosmanpaşa Univ Tıp Fak Derg. 2014;6(4):281-294.
42. Topal K, Aybek H, Kara CO, Büke A, Aybek Z. PAÜ Tıp Fakültesi dönem
ı öğrencilerine 2006-2007 eğitim ve öğretim yılında uygulanan çoktan seçmeli sınavların madde ve test analizleri. Pamukkale Tıp Derg.
2008;1(3):120-126.
43. Tiemeier AM, Stacy ZA, Burke JM. Using multiple choice questions
written at various Bloom’s taxonomy levels to evaluate student performance across a therapeutics sequence. Innov Pharm. 2011;2(2):1-
11. [CrossRef]
44. Simkin MG, Kuechler WL. Multiple-choice tests and student understanding: What is the connection? Decis Sci J Innov Educ.
2005;3(1):73-98. [CrossRe
Yılmaz Koğar E. Sağlık Alanından Örneklerle Çoktan Seçmeli Maddelerin Üst Düzey Düşünme Becerilerine Yönelik Tasarlanması için Öneriler. Journal of Nursology. June 2022;25(2):117-125. doi:10.5152/JANHS.2022.875827