Publication Date
| In 2026 | 0 |
| Since 2025 | 28 |
| Since 2022 (last 5 years) | 146 |
| Since 2017 (last 10 years) | 359 |
| Since 2007 (last 20 years) | 586 |
Descriptor
| Multiple Choice Tests | 1156 |
| Test Items | 1156 |
| Test Construction | 416 |
| Foreign Countries | 337 |
| Difficulty Level | 298 |
| Test Format | 260 |
| Item Analysis | 244 |
| Item Response Theory | 177 |
| Test Reliability | 172 |
| Higher Education | 162 |
| Test Validity | 161 |
| More ▼ | |
Source
Author
| Haladyna, Thomas M. | 14 |
| Plake, Barbara S. | 8 |
| Samejima, Fumiko | 8 |
| Downing, Steven M. | 7 |
| Bennett, Randy Elliot | 6 |
| Cheek, Jimmy G. | 6 |
| Huntley, Renee M. | 6 |
| Katz, Irvin R. | 6 |
| Kim, Sooyeon | 6 |
| McGhee, Max B. | 6 |
| Suh, Youngsuk | 6 |
| More ▼ | |
Publication Type
Education Level
Audience
| Practitioners | 40 |
| Students | 30 |
| Teachers | 28 |
| Researchers | 26 |
| Administrators | 5 |
| Counselors | 1 |
Location
| Canada | 62 |
| Australia | 37 |
| Turkey | 29 |
| Indonesia | 22 |
| Germany | 14 |
| Iran | 11 |
| Nigeria | 11 |
| Malaysia | 10 |
| China | 9 |
| Taiwan | 9 |
| United Kingdom | 9 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 4 |
| National Defense Education Act | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
| Does not meet standards | 1 |
Ute Mertens; Marlit A. Lindner – Journal of Computer Assisted Learning, 2025
Background: Educational assessments increasingly shift towards computer-based formats. Many studies have explored how different types of automated feedback affect learning. However, few studies have investigated how digital performance feedback affects test takers' ratings of affective-motivational reactions during a testing session. Method: In…
Descriptors: Educational Assessment, Computer Assisted Testing, Automation, Feedback (Response)
Andrew M. Olney – Grantee Submission, 2023
Multiple choice questions are traditionally expensive to produce. Recent advances in large language models (LLMs) have led to fine-tuned LLMs that generate questions competitive with human-authored questions. However, the relative capabilities of ChatGPT-family models have not yet been established for this task. We present a carefully-controlled…
Descriptors: Test Construction, Multiple Choice Tests, Test Items, Algorithms
Testing Anatomy: Dissecting Spatial and Non-Spatial Knowledge in Multiple-Choice Question Assessment
Julie Dickson; Darren J. Shaw; Andrew Gardiner; Susan Rhind – Anatomical Sciences Education, 2024
Limited research has been conducted on the spatial ability of veterinary students and how this is evaluated within anatomy assessments. This study describes the creation and evaluation of a split design multiple-choice question (MCQ) assessment (totaling 30 questions divided into 15 non-spatial MCQs and 15 spatial MCQs). Two cohorts were tested,…
Descriptors: Anatomy, Spatial Ability, Multiple Choice Tests, Factor Analysis
Chukwudi Ikwueze – Educational Research Quarterly, 2024
The U.S. Department of Education advocates strongly using assessment of student learning as a means of maintaining high standards of education. Instructors use mostly multiple choice and discussion questions as assessment tools. The aim of this study was to evaluate the effectiveness of multiple-choice and discussion questions and participant…
Descriptors: Economics Education, Introductory Courses, Multiple Choice Tests, Student Evaluation
Chi-Chen Chen; Chia-Wei Tang; Kuan-Yu Jin – Large-scale Assessments in Education, 2024
Internet-related issues have influenced how civic knowledge is educated and measured. The International Civic and Citizenship Education Study (ICCS) is a well-known large-scale assessment concerning how civic knowledge is educated and measured globally. Regardless of the emerging roles of internet access and usage, the influences of internet…
Descriptors: Internet, Civics, Citizenship Education, Student Surveys
Brennan, Robert L.; Kim, Stella Y.; Lee, Won-Chan – Educational and Psychological Measurement, 2022
This article extends multivariate generalizability theory (MGT) to tests with different random-effects designs for each level of a fixed facet. There are numerous situations in which the design of a test and the resulting data structure are not definable by a single design. One example is mixed-format tests that are composed of multiple-choice and…
Descriptors: Multivariate Analysis, Generalizability Theory, Multiple Choice Tests, Test Construction
Davison, Mark L.; Davenport, Ernest C., Jr.; Jia, Hao; Seipel, Ben; Carlson, Sarah E. – Grantee Submission, 2022
A regression model of predictor trade-offs is described. Each regression parameter equals the expected change in Y obtained by trading 1 point from one predictor to a second predictor. The model applies to predictor variables that sum to a constant T for all observations; for example, proportions summing to T=1.0 or percentages summing to T=100…
Descriptors: Regression (Statistics), Prediction, Predictor Variables, Models
Baghaei, Purya; Christensen, Karl Bang – Language Testing, 2023
C-tests are gap-filling tests mainly used as rough and economical measures of second-language proficiency for placement and research purposes. A C-test usually consists of several short independent passages where the second half of every other word is deleted. Owing to their interdependent structure, C-test items violate the local independence…
Descriptors: Item Response Theory, Language Tests, Language Proficiency, Second Language Learning
Valentina Albano; Donatella Firmani; Luigi Laura; Jerin George Mathew; Anna Lucia Paoletti; Irene Torrente – Journal of Learning Analytics, 2023
Multiple-choice questions (MCQs) are widely used in educational assessments and professional certification exams. Managing large repositories of MCQs, however, poses several challenges due to the high volume of questions and the need to maintain their quality and relevance over time. One of these challenges is the presence of questions that…
Descriptors: Natural Language Processing, Multiple Choice Tests, Test Items, Item Analysis
van den Broek, Gesa S. E.; Gerritsen, Suzanne L.; Oomen, Iris T. J.; Velthoven, Eva; van Boxtel, Femke H. J.; Kester, Liesbeth; van Gog, Tamara – Journal of Educational Psychology, 2023
Multiple-choice questions (MCQs) are popular in vocabulary software because they can be scored automatically and are compatible with many input devices (e.g., touchscreens). Answering MCQs is beneficial for learning, especially when learners retrieve knowledge from memory to evaluate plausible answer alternatives. However, such retrieval may not…
Descriptors: Multiple Choice Tests, Vocabulary Development, Test Format, Cues
Kunal Sareen – Innovations in Education and Teaching International, 2024
This study examines the proficiency of Chat GPT, an AI language model, in answering questions on the Situational Judgement Test (SJT), a widely used assessment tool for evaluating the fundamental competencies of medical graduates in the UK. A total of 252 SJT questions from the "Oxford Assess and Progress: Situational Judgement" Test…
Descriptors: Ethics, Decision Making, Artificial Intelligence, Computer Software
Guo, Wenjing; Wind, Stefanie A. – Journal of Educational Measurement, 2021
The use of mixed-format tests made up of multiple-choice (MC) items and constructed response (CR) items is popular in large-scale testing programs, including the National Assessment of Educational Progress (NAEP) and many district- and state-level assessments in the United States. Rater effects, or raters' scoring tendencies that result in…
Descriptors: Test Format, Multiple Choice Tests, Scoring, Test Items
Roger Young; Emily Courtney; Alexander Kah; Mariah Wilkerson; Yi-Hsin Chen – Teaching of Psychology, 2025
Background: Multiple-choice item (MCI) assessments are burdensome for instructors to develop. Artificial intelligence (AI, e.g., ChatGPT) can streamline the process without sacrificing quality. The quality of AI-generated MCIs and human experts is comparable. However, whether the quality of AI-generated MCIs is equally good across various domain-…
Descriptors: Item Response Theory, Multiple Choice Tests, Psychology, Textbooks
Mead, Alan D.; Zhou, Chenxuan – Journal of Applied Testing Technology, 2022
This study fit a Naïve Bayesian classifier to the words of exam items to predict the Bloom's taxonomy level of the items. We addressed five research questions, showing that reasonably good prediction of Bloom's level was possible, but accuracy varies across levels. In our study, performance for Level 2 was poor (Level 2 items were misclassified…
Descriptors: Artificial Intelligence, Prediction, Taxonomy, Natural Language Processing
Marli Crabtree; Kenneth L. Thompson; Ellen M. Robertson – HAPS Educator, 2024
Research has suggested that changing one's answer on multiple-choice examinations is more likely to lead to positive academic outcomes. This study aimed to further understand the relationship between changing answer selections and item attributes, student performance, and time within a population of 158 first-year medical students enrolled in a…
Descriptors: Anatomy, Science Tests, Medical Students, Medical Education

Peer reviewed
Direct link
