Publication Date
| In 2026 | 0 |
| Since 2025 | 15 |
| Since 2022 (last 5 years) | 56 |
| Since 2017 (last 10 years) | 147 |
| Since 2007 (last 20 years) | 261 |
Descriptor
| Difficulty Level | 426 |
| Multiple Choice Tests | 426 |
| Test Items | 298 |
| Test Construction | 134 |
| Foreign Countries | 131 |
| Item Analysis | 103 |
| Test Format | 96 |
| Test Reliability | 85 |
| Item Response Theory | 79 |
| Test Validity | 74 |
| Higher Education | 70 |
| More ▼ | |
Source
Author
| Tindal, Gerald | 6 |
| Alonzo, Julie | 5 |
| DeBoer, George E. | 5 |
| Herrmann-Abell, Cari F. | 5 |
| Plake, Barbara S. | 5 |
| Cizek, Gregory J. | 4 |
| Huntley, Renee M. | 4 |
| Katz, Irvin R. | 4 |
| Tollefson, Nona | 4 |
| Anderson, Paul S. | 3 |
| Andrich, David | 3 |
| More ▼ | |
Publication Type
Education Level
| Higher Education | 106 |
| Postsecondary Education | 86 |
| Secondary Education | 64 |
| Elementary Education | 45 |
| Middle Schools | 30 |
| High Schools | 24 |
| Junior High Schools | 19 |
| Intermediate Grades | 17 |
| Grade 6 | 13 |
| Grade 7 | 13 |
| Grade 5 | 12 |
| More ▼ | |
Audience
| Researchers | 10 |
| Teachers | 2 |
| Administrators | 1 |
| Practitioners | 1 |
Location
| Turkey | 14 |
| Indonesia | 10 |
| Australia | 8 |
| Canada | 8 |
| Germany | 8 |
| Nigeria | 7 |
| Taiwan | 6 |
| Jordan | 5 |
| Netherlands | 5 |
| California | 4 |
| Malaysia | 4 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 2 |
| Elementary and Secondary… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Arandha May Rachmawati; Agus Widyantoro – English Language Teaching Educational Journal, 2025
This study aims to evaluate the quality of English reading comprehension test instruments used in informal learning, especially as English literacy tests. With a quantitative approach, the analysis was carried out using the Rasch model through the Quest program on 30 multiple-choice questions given to 30 grade IX students from informal educational…
Descriptors: Item Response Theory, Reading Tests, Reading Comprehension, English (Second Language)
Jin, Kuan-Yu; Siu, Wai-Lok; Huang, Xiaoting – Journal of Educational Measurement, 2022
Multiple-choice (MC) items are widely used in educational tests. Distractor analysis, an important procedure for checking the utility of response options within an MC item, can be readily implemented in the framework of item response theory (IRT). Although random guessing is a popular behavior of test-takers when answering MC items, none of the…
Descriptors: Guessing (Tests), Multiple Choice Tests, Item Response Theory, Attention
Thayaamol Upapong; Apantee Poonputta – Educational Process: International Journal, 2025
Background/purpose: The purposes of this research are to develop a reliable and valid assessment tool for measuring systems thinking skills in upper primary students in Thailand and to establish a normative criterion for evaluating their systems thinking abilities based on educational standards. Materials/methods: The study followed a three-phase…
Descriptors: Thinking Skills, Elementary School Students, Measures (Individuals), Foreign Countries
Eka Febri Zulissetiana; Muhammad Irfannuddin; Siti Sarahdeaz Fazzaura Putri; Syifa Alkaf; Susilawati Susilawati; Jihan Marshanda; Ra Fadila Septiany; Hasyimiah Az-Zahra; Robert G. Carroll – Advances in Physiology Education, 2024
Complex subjects such as physiology can be challenging for students to learn. These challenges are not uncommon in implementing the learning process in physiology and affect learning outcomes. Dramatization is an interactive and effective method to improve learning outcomes. In a project designed by senior medical students, junior medical students…
Descriptors: Drama, Teaching Methods, Physiology, Science Instruction
Kuo-Zheng Feng – Language Testing in Asia, 2024
This study addressed a gap in existing research on Multiple-Choice (MC) cloze tests by focusing on the learners' perspective, specifically examining the difficulties faced by vocational high school students (VHSs). A nationwide sample of 293 VHSs participated, providing both quantitative and qualitative data through a self-developed questionnaire.…
Descriptors: Language Tests, Multiple Choice Tests, Cloze Procedure, Student Attitudes
van den Broek, Gesa S. E.; Gerritsen, Suzanne L.; Oomen, Iris T. J.; Velthoven, Eva; van Boxtel, Femke H. J.; Kester, Liesbeth; van Gog, Tamara – Journal of Educational Psychology, 2023
Multiple-choice questions (MCQs) are popular in vocabulary software because they can be scored automatically and are compatible with many input devices (e.g., touchscreens). Answering MCQs is beneficial for learning, especially when learners retrieve knowledge from memory to evaluate plausible answer alternatives. However, such retrieval may not…
Descriptors: Multiple Choice Tests, Vocabulary Development, Test Format, Cues
Roger Young; Emily Courtney; Alexander Kah; Mariah Wilkerson; Yi-Hsin Chen – Teaching of Psychology, 2025
Background: Multiple-choice item (MCI) assessments are burdensome for instructors to develop. Artificial intelligence (AI, e.g., ChatGPT) can streamline the process without sacrificing quality. The quality of AI-generated MCIs and human experts is comparable. However, whether the quality of AI-generated MCIs is equally good across various domain-…
Descriptors: Item Response Theory, Multiple Choice Tests, Psychology, Textbooks
Anatri Desstya; Ika Candra Sayekti; Muhammad Abduh; Sukartono – Journal of Turkish Science Education, 2025
This study aimed to develop a standardised instrument for diagnosing science misconceptions in primary school children. Following a developmental research approach using the 4-D model (Define, Design, Develop, Disseminate), 100 four-tier multiple choice items were constructed. Content validity was established through expert evaluation by six…
Descriptors: Test Construction, Science Tests, Science Instruction, Diagnostic Tests
Liu, Chunyan; Jurich, Daniel; Morrison, Carol; Grabovsky, Irina – Applied Measurement in Education, 2021
The existence of outliers in the anchor items can be detrimental to the estimation of examinee ability and undermine the validity of score interpretation across forms. However, in practice, anchor item performance can become distorted due to various reasons. This study compares the performance of modified "INFIT" and "OUTFIT"…
Descriptors: Equated Scores, Test Items, Item Response Theory, Difficulty Level
Kevser Arslan; Asli Görgülü Ari – Shanlax International Journal of Education, 2024
This study aimed to develop a valid and reliable multiple-choice achievement test for the subject area of ecology. The study was conducted within the framework of exploratory sequential design based on mixed research methods, and the study group consisted of a total of 250 middle school students studying at the sixth and seventh grade level. In…
Descriptors: Ecology, Science Tests, Test Construction, Multiple Choice Tests
DeCarlo, Lawrence T. – Journal of Educational Measurement, 2023
A conceptualization of multiple-choice exams in terms of signal detection theory (SDT) leads to simple measures of item difficulty and item discrimination that are closely related to, but also distinct from, those used in classical item analysis (CIA). The theory defines a "true split," depending on whether or not examinees know an item,…
Descriptors: Multiple Choice Tests, Test Items, Item Analysis, Test Wiseness
Alicia A. Stoltenberg – ProQuest LLC, 2024
Multiple-select multiple-choice items, or multiple-choice items with more than one correct answer, are used to quickly assess content on standardized assessments. Because there are multiple keys to these item types, there are also multiple ways to score student responses to these items. The purpose of this study was to investigate how changing the…
Descriptors: Scoring, Evaluation Methods, Multiple Choice Tests, Standardized Tests
Al-zboon, Habis Saad; Alrekebat, Amjad Farhan – International Journal of Higher Education, 2021
This study aims at identifying the effect of multiple-choice test items' difficulty degree on the reliability coefficient and the standard error of measurement depending on the item response theory IRT. To achieve the objectives of the study, (WinGen3) software was used to generate the IRT parameters (difficulty, discrimination, guessing) for four…
Descriptors: Multiple Choice Tests, Test Items, Difficulty Level, Error of Measurement
Herrmann-Abell, Cari F.; Hardcastle, Joseph; DeBoer, George E. – Grantee Submission, 2022
As implementation of the "Next Generation Science Standards" moves forward, there is a need for new assessments that can measure students' integrated three-dimensional science learning. The National Research Council has suggested that these assessments be multicomponent tasks that utilize a combination of item formats including…
Descriptors: Multiple Choice Tests, Conditioning, Test Items, Item Response Theory
Büsra Kilinç; Mehmet Diyaddin Yasar – Science Insights Education Frontiers, 2024
In this study, it was aimed to develop an achievement test taking into account the subject acquisitions of the sound and properties unit in the sixth-grade science course. In the test development phase, firstly, literature review for the study was conducted. Then, 30 multiple choice questions in align with the subject acquisition in the 2018…
Descriptors: Science Tests, Test Construction, Grade 6, Science Instruction

Peer reviewed
Direct link
