Publication Date
| In 2026 | 0 |
| Since 2025 | 2 |
| Since 2022 (last 5 years) | 16 |
| Since 2017 (last 10 years) | 44 |
Descriptor
| Multiple Choice Tests | 44 |
| Scoring | 44 |
| Test Items | 20 |
| Foreign Countries | 18 |
| Science Tests | 10 |
| College Students | 9 |
| Computer Assisted Testing | 9 |
| Item Response Theory | 9 |
| Test Reliability | 9 |
| Evaluation Methods | 8 |
| Language Tests | 8 |
| More ▼ | |
Source
Author
Publication Type
| Reports - Research | 41 |
| Journal Articles | 40 |
| Numerical/Quantitative Data | 2 |
| Speeches/Meeting Papers | 2 |
| Dissertations/Theses -… | 1 |
| Reports - Descriptive | 1 |
| Reports - Evaluative | 1 |
| Tests/Questionnaires | 1 |
Education Level
| Higher Education | 17 |
| Postsecondary Education | 17 |
| Secondary Education | 13 |
| Elementary Education | 8 |
| Junior High Schools | 4 |
| Middle Schools | 4 |
| Grade 8 | 3 |
| High Schools | 3 |
| Grade 10 | 1 |
| Grade 11 | 1 |
| Grade 12 | 1 |
| More ▼ | |
Audience
Location
| Turkey | 4 |
| China | 3 |
| United States | 2 |
| Australia | 1 |
| Europe | 1 |
| Indonesia | 1 |
| Iran | 1 |
| Malaysia | 1 |
| Nigeria | 1 |
| Portugal | 1 |
| Russia | 1 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
| National Assessment of… | 3 |
| English Proficiency Test | 1 |
| Flesch Kincaid Grade Level… | 1 |
What Works Clearinghouse Rating
Stefanie A. Wind; Yuan Ge – Measurement: Interdisciplinary Research and Perspectives, 2024
Mixed-format assessments made up of multiple-choice (MC) items and constructed response (CR) items that are scored using rater judgments include unique psychometric considerations. When these item types are combined to estimate examinee achievement, information about the psychometric quality of each component can depend on that of the other. For…
Descriptors: Interrater Reliability, Test Bias, Multiple Choice Tests, Responses
Sarah Alahmadi; Christine E. DeMars – Applied Measurement in Education, 2024
Large-scale educational assessments are sometimes considered low-stakes, increasing the possibility of confounding true performance level with low motivation. These concerns are amplified in remote testing conditions. To remove the effects of low effort levels in responses observed in remote low-stakes testing, several motivation filtering methods…
Descriptors: Multiple Choice Tests, Item Response Theory, College Students, Scores
Filipe Leite-Mendes; Luis Delgado; Amelia Ferreira; Milton Severo – Advances in Health Sciences Education, 2024
Given the high prevalence of multiple-choice examinations with formula scoring in medical training, several studies have tried to identify other factors in addition to the degree of knowledge of students which influence their response patterns. This study aims to measure the effect of students' attitude towards risk and ambiguity on their number…
Descriptors: Foreign Countries, Medical Schools, Medical Students, Student Attitudes
Herwin, Herwin; Pristiwaluyo, Triyanto; Ruslan, Ruslan; Dahalan, Shakila Che – Cypriot Journal of Educational Sciences, 2022
The application of multiple-choice tests often does not consider the scoring technique and the number of choices. The study aims at describing the effect of the scoring technique and numerous options towards the reliability of multiple-choice objective tests on social subjects in elementary school. The study is quantitative research with…
Descriptors: Scoring, Multiple Choice Tests, Test Reliability, Elementary School Students
Wind, Stefanie A.; Ge, Yuan – Educational and Psychological Measurement, 2021
Practical constraints in rater-mediated assessments limit the availability of complete data. Instead, most scoring procedures include one or two ratings for each performance, with overlapping performances across raters or linking sets of multiple-choice items to facilitate model estimation. These incomplete scoring designs present challenges for…
Descriptors: Evaluators, Scoring, Data Collection, Design
Kunal Sareen – Innovations in Education and Teaching International, 2024
This study examines the proficiency of Chat GPT, an AI language model, in answering questions on the Situational Judgement Test (SJT), a widely used assessment tool for evaluating the fundamental competencies of medical graduates in the UK. A total of 252 SJT questions from the "Oxford Assess and Progress: Situational Judgement" Test…
Descriptors: Ethics, Decision Making, Artificial Intelligence, Computer Software
Guo, Wenjing; Wind, Stefanie A. – Journal of Educational Measurement, 2021
The use of mixed-format tests made up of multiple-choice (MC) items and constructed response (CR) items is popular in large-scale testing programs, including the National Assessment of Educational Progress (NAEP) and many district- and state-level assessments in the United States. Rater effects, or raters' scoring tendencies that result in…
Descriptors: Test Format, Multiple Choice Tests, Scoring, Test Items
Kristen L. Murphy; David G. Schreurs; Melonie A. Teichert; Cynthia J. Luxford; Jaclyn M. Trate; Jordan T. Harshmann; Jamie L. Schneider – Chemistry Education Research and Practice, 2024
Providing students with feedback on their performance is a critical part of enhancing student learning in chemistry and is often integrated into homework assignments, quizzes, and exams. However, not all feedback is created equal, and the type of feedback the student receives can dramatically alter the utility of the feedback to reinforce correct…
Descriptors: Student Evaluation, Feedback (Response), Science Education, Introductory Courses
Zhiqiang Yang; Chengyuan Yu – Asia Pacific Education Review, 2025
This study investigated the test fairness of the translation section of a large-scale English test in China by examining its Differential Test Functioning (DTF) and Differential Item Functioning (DIF) across gender and major. Regarding DTF, the entire translation section exhibits partial strong measurement invariance across female and male…
Descriptors: Multiple Choice Tests, Test Items, Scoring, Translation
Alicia A. Stoltenberg – ProQuest LLC, 2024
Multiple-select multiple-choice items, or multiple-choice items with more than one correct answer, are used to quickly assess content on standardized assessments. Because there are multiple keys to these item types, there are also multiple ways to score student responses to these items. The purpose of this study was to investigate how changing the…
Descriptors: Scoring, Evaluation Methods, Multiple Choice Tests, Standardized Tests
Güntay Tasçi – Science Insights Education Frontiers, 2024
The present study has aimed to develop and validate a protein concept inventory (PCI) consisting of 25 multiple-choice (MC) questions to assess students' understanding of protein, which is a fundamental concept across different biology disciplines. The development process of the PCI involved a literature review to identify protein-related content,…
Descriptors: Science Instruction, Science Tests, Multiple Choice Tests, Biology
Congning Ni; Bhashithe Abeysinghe; Juanita Hicks – International Electronic Journal of Elementary Education, 2025
The National Assessment of Educational Progress (NAEP), often referred to as The Nation's Report Card, offers a window into the state of U.S. K-12 education system. Since 2017, NAEP has transitioned to digital assessments, opening new research opportunities that were previously impossible. Process data tracks students' interactions with the…
Descriptors: Reaction Time, Multiple Choice Tests, Behavior Change, National Competency Tests
Zhai, Xiaoming; Li, Min – International Journal of Science Education, 2021
This study provides a partial-credit scoring (PCS) approach to awarding students' performance on multiple-choice items in science education. The approach is built on "fundamental ideas," the critical pieces of students' understanding and knowledge to solve science problems. We link each option of the items to several specific fundamental…
Descriptors: Scoring, Multiple Choice Tests, Science Tests, Test Items
Çinar, Ayse; Ince, Elif; Gezer, Murat; Yilmaz, Özgür – Education and Information Technologies, 2020
Worldwide, open-ended questions that require short answers have been used in many exams in fields of science, such as the International Student Assessment Program (PISA), the International Science and Maths Trends Research (TIMSS). However, multiple-choice questions are used for many exams at the national level in Turkey, especially high school…
Descriptors: Foreign Countries, Computer Assisted Testing, Artificial Intelligence, Grading
Slepkov, Aaron D.; Godfrey, Alan T. K. – Applied Measurement in Education, 2019
The answer-until-correct (AUC) method of multiple-choice (MC) testing involves test respondents making selections until the keyed answer is identified. Despite attendant benefits that include improved learning, broad student adoption, and facile administration of partial credit, the use of AUC methods for classroom testing has been extremely…
Descriptors: Multiple Choice Tests, Test Items, Test Reliability, Scores

Peer reviewed
Direct link
