Publication Date
| In 2026 | 0 |
| Since 2025 | 26 |
| Since 2022 (last 5 years) | 144 |
| Since 2017 (last 10 years) | 357 |
| Since 2007 (last 20 years) | 584 |
Descriptor
| Multiple Choice Tests | 1154 |
| Test Items | 1154 |
| Test Construction | 414 |
| Foreign Countries | 336 |
| Difficulty Level | 298 |
| Test Format | 260 |
| Item Analysis | 244 |
| Item Response Theory | 177 |
| Test Reliability | 172 |
| Higher Education | 162 |
| Test Validity | 161 |
| More ▼ | |
Source
Author
| Haladyna, Thomas M. | 14 |
| Plake, Barbara S. | 8 |
| Samejima, Fumiko | 8 |
| Downing, Steven M. | 7 |
| Bennett, Randy Elliot | 6 |
| Cheek, Jimmy G. | 6 |
| Huntley, Renee M. | 6 |
| Katz, Irvin R. | 6 |
| Kim, Sooyeon | 6 |
| McGhee, Max B. | 6 |
| Suh, Youngsuk | 6 |
| More ▼ | |
Publication Type
Education Level
Audience
| Practitioners | 40 |
| Students | 30 |
| Teachers | 28 |
| Researchers | 26 |
| Administrators | 5 |
| Counselors | 1 |
Location
| Canada | 62 |
| Australia | 37 |
| Turkey | 29 |
| Indonesia | 22 |
| Germany | 14 |
| Iran | 11 |
| Nigeria | 11 |
| Malaysia | 10 |
| China | 9 |
| Taiwan | 9 |
| United Kingdom | 9 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 4 |
| National Defense Education Act | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
| Does not meet standards | 1 |
Rao, Dhawaleswar; Saha, Sujan Kumar – IEEE Transactions on Learning Technologies, 2020
Automatic multiple choice question (MCQ) generation from a text is a popular research area. MCQs are widely accepted for large-scale assessment in various domains and applications. However, manual generation of MCQs is expensive and time-consuming. Therefore, researchers have been attracted toward automatic MCQ generation since the late 90's.…
Descriptors: Multiple Choice Tests, Test Construction, Automation, Computer Software
Bulut, Okan; Bulut, Hatice Cigdem; Cormier, Damien C.; Ilgun Dibek, Munevver; Sahin Kursad, Merve – Educational Assessment, 2023
Some statewide testing programs allow students to receive corrective feedback and revise their answers during testing. Despite its pedagogical benefits, the effects of providing revision opportunities remain unknown in the context of alternate assessments. Therefore, this study examined student data from a large-scale alternate assessment that…
Descriptors: Error Correction, Alternative Assessment, Feedback (Response), Multiple Choice Tests
Deribo, Tobias; Goldhammer, Frank; Kroehne, Ulf – Educational and Psychological Measurement, 2023
As researchers in the social sciences, we are often interested in studying not directly observable constructs through assessments and questionnaires. But even in a well-designed and well-implemented study, rapid-guessing behavior may occur. Under rapid-guessing behavior, a task is skimmed shortly but not read and engaged with in-depth. Hence, a…
Descriptors: Reaction Time, Guessing (Tests), Behavior Patterns, Bias
Opstad, Leiv – Athens Journal of Education, 2021
The discussion of whether multiple-choice questions can replace the traditional exam with essays and constructed questions in introductory courses has just started in Norway. There is not an easy answer. The findings depend on the pattern of the questions. Therefore, one must be careful in drawing conclusions. In this research, one will explore a…
Descriptors: Multiple Choice Tests, Essay Tests, Introductory Courses, Foreign Countries
Mustafa, Nazahiyah; Khairani, Ahmad Zamri; Ishak, Nor Asniza – International Journal of Evaluation and Research in Education, 2021
This study aimed to calibrate the test items of science process skills used as a test at primary school students to provide information on the difficulty of each item. Data were collected from 128 standard five students in a primary school in Penang. The test was given in multiple-choice as many as 40 items consisting of 33 items partial credit…
Descriptors: Elementary School Students, Elementary School Science, Science Process Skills, Foreign Countries
Shear, Benjamin R. – Journal of Educational Measurement, 2023
Large-scale standardized tests are regularly used to measure student achievement overall and for student subgroups. These uses assume tests provide comparable measures of outcomes across student subgroups, but prior research suggests score comparisons across gender groups may be complicated by the type of test items used. This paper presents…
Descriptors: Gender Bias, Item Analysis, Test Items, Achievement Tests
Slepkov, Aaron D.; Godfrey, Alan T. K. – Applied Measurement in Education, 2019
The answer-until-correct (AUC) method of multiple-choice (MC) testing involves test respondents making selections until the keyed answer is identified. Despite attendant benefits that include improved learning, broad student adoption, and facile administration of partial credit, the use of AUC methods for classroom testing has been extremely…
Descriptors: Multiple Choice Tests, Test Items, Test Reliability, Scores
Burfitt, Joan – Issues in Educational Research, 2019
During the development of a test of multiple-choice items, cognitive interviews were conducted with students from lower secondary school. The purpose of these interviews was to confirm that the prospective respondents' interpretation of mathematics test items was consistent with the interpretation intended by the item writer. The conversations…
Descriptors: Interviews, Multiple Choice Tests, Mathematics Tests, Secondary School Students
Setiawan, Johan; Sudrajat, Ajat; Aman; Kumalasari, Dyah – International Journal of Evaluation and Research in Education, 2021
This study aimed to: (1) produce higher order thinking skill (HOTS) assessment instruments in learning Indonesian history; (2) know the validity of HOTS assessment instruments in learning Indonesian history; and (3) find out the characteristics of HOTS questions in learning Indonesian history. This study employed the research and development…
Descriptors: Foreign Countries, History Instruction, Thinking Skills, Test Construction
Budi Waluyo; Ali Zahabi; Luksika Ruangsung – rEFLections, 2024
The increasing popularity of the Common European Framework of Reference (CEFR) in non-native English-speaking countries has generated a demand for concrete examples in the creation of CEFR-based tests that assess the four main English skills. In response, this research endeavors to provide insight into the development and validation of a…
Descriptors: Language Tests, Language Proficiency, Undergraduate Students, Language Skills
Ali, Ikram; Haral, Muhammad Nouman; Tahira, Fatima; Ali, Mirza; Imran, Aqeel – Education and Urban Society, 2020
Foundation of quality examination is based on the key features of validity and reliability of question papers. Question papers of examination boards in Pakistan--usually these features as contents of a question paper--are written by a single paper setter. To address these issues, Federal Board of Intermediate and Secondary Education (FBISE)…
Descriptors: Foreign Countries, Multiple Choice Tests, Test Items, Difficulty Level
Shin, Jinnie; Bulut, Okan; Gierl, Mark J. – Journal of Experimental Education, 2020
The arrangement of response options in multiple-choice (MC) items, especially the location of the most attractive distractor, is considered critical in constructing high-quality MC items. In the current study, a sample of 496 undergraduate students taking an educational assessment course was given three test forms consisting of the same items but…
Descriptors: Foreign Countries, Undergraduate Students, Multiple Choice Tests, Item Response Theory
Saepuzaman, Duden; Istiyono, Edi; Haryanto – Pegem Journal of Education and Instruction, 2022
HOTS is one part of the skills that need to be developed in the 21st Century . This study aims to determine the characteristics of the Fundamental Physics Higher-order Thinking Skill (FundPhysHOTS) test for prospective physics teachers using Item Response Theory (IRT) analysis. This study uses a quantitative approach. 254 prospective physics…
Descriptors: Thinking Skills, Physics, Science Process Skills, Cognitive Tests
Tomkowicz, Joanna; Kim, Dong-In; Wan, Ping – Online Submission, 2022
In this study we evaluated the stability of item parameters and student scores, using the pre-equated (pre-pandemic) parameters from Spring 2019 and post-equated (post-pandemic) parameters from Spring 2021 in two calibration and equating designs related to item parameter treatment: re-estimating all anchor parameters (Design 1) and holding the…
Descriptors: Equated Scores, Test Items, Evaluation Methods, Pandemics
Azevedo, Jose Manuel; Oliveira, Ema P.; Beites, Patrícia Damas – International Journal of Information and Learning Technology, 2019
Purpose: The purpose of this paper is to find appropriate forms of analysis of multiple-choice questions (MCQ) to obtain an assessment method, as fair as possible, for the students. The authors intend to ascertain if it is possible to control the quality of the MCQ contained in a bank of questions, implemented in Moodle, presenting some evidence…
Descriptors: Learning Analytics, Multiple Choice Tests, Test Theory, Item Response Theory

Peer reviewed
Direct link
