Publication Date
| In 2026 | 0 |
| Since 2025 | 10 |
| Since 2022 (last 5 years) | 40 |
| Since 2017 (last 10 years) | 118 |
| Since 2007 (last 20 years) | 211 |
Descriptor
| Multiple Choice Tests | 532 |
| Test Reliability | 532 |
| Test Validity | 302 |
| Test Construction | 238 |
| Test Items | 172 |
| Foreign Countries | 114 |
| Item Analysis | 101 |
| Higher Education | 90 |
| Difficulty Level | 85 |
| Guessing (Tests) | 74 |
| Scoring | 69 |
| More ▼ | |
Source
Author
| Ebel, Robert L. | 10 |
| Frary, Robert B. | 9 |
| Alonzo, Julie | 7 |
| Frisbie, David A. | 6 |
| Irvin, P. Shawn | 6 |
| Lai, Cheng-Fei | 6 |
| Park, Bitnara Jasmine | 6 |
| Tindal, Gerald | 6 |
| Wilcox, Rand R. | 5 |
| Albanese, Mark A. | 4 |
| Biancarosa, Gina | 4 |
| More ▼ | |
Publication Type
Education Level
Audience
| Researchers | 11 |
| Practitioners | 8 |
| Teachers | 5 |
Location
| Indonesia | 17 |
| Turkey | 17 |
| Germany | 8 |
| Iran | 8 |
| Canada | 6 |
| Malaysia | 4 |
| Nigeria | 4 |
| Australia | 3 |
| Florida | 3 |
| Japan | 3 |
| Pakistan | 3 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Peer reviewedPyrczak, Fred – Journal of Reading, 1977
Students know more than they think they know, so guessing gives better scores even when there's a penalty for errors. (JM)
Descriptors: Guessing (Tests), Multiple Choice Tests, Reading Research, Reading Tests
Peer reviewedTollefson, Nona – Educational and Psychological Measurement, 1987
This study compared the item difficulty, item discrimination, and test reliability of three forms of multiple-choice items: (1) one correct answer; (2) "none of the above" as a foil; and (3) "none of the above" as the correct answer. Twelve items in the three formats were administered in a college statistics examination. (BS)
Descriptors: Difficulty Level, Higher Education, Item Analysis, Multiple Choice Tests
Peer reviewedZimmerman, Donald W. – Educational and Psychological Measurement, 1985
A computer program simulated guessing on multiple-choice test items and calculated deviation IQ's from observed scores which contained a guessing component. Extensive variability in deviation IQ's due entirely to chance was found. (Author/LMO)
Descriptors: Computer Simulation, Error of Measurement, Guessing (Tests), Intelligence Quotient
Peer reviewedWilcox, Rand R. – Psychometrika, 1983
A procedure for determining the reliability of an examinee knowing k out of n possible multiple choice items given his or her performance on those items is presented. Also, a scoring procedure for determining which items an examinee knows is presented. (Author/JKS)
Descriptors: Item Analysis, Latent Trait Theory, Measurement Techniques, Multiple Choice Tests
Peer reviewedBoard, Cynthia; Whitney, Douglas R. – Journal of Educational Measurement, 1972
For the principles studied here, poor item-writing practices serve to obscure (or attentuate) differences between good and poor students. (Authors)
Descriptors: College Students, Item Analysis, Multiple Choice Tests, Test Construction
Rippey, Robert M. – J Educ Meas, 1970
Descriptors: Multiple Choice Tests, Prediction, Probability, Scoring
Peer reviewedHansen, Richard – Journal of Educational Measurement, 1971
The relationship between certain personality variables and the degree to which examines display certainty in their responses was investigated. (Author)
Descriptors: Guessing (Tests), Individual Characteristics, Multiple Choice Tests, Personality Assessment
Peer reviewedBornstein, Harry; Chamberlain, Karen – American Educational Research Journal, 1970
Descriptors: Achievement Tests, Language Handicaps, Multiple Choice Tests, Syntax
Peer reviewedFrary, Robert B. – Journal of Educational Statistics, 1982
Six different approaches to scoring test data, including number right, correction for guessing, and answer-until-correct, were investigated using Monte Carlo techniques. Modes permitting multiple response showed higher internal consistency, but there was little difference among modes for a validity measure. (JKS)
Descriptors: Guessing (Tests), Measurement Techniques, Multiple Choice Tests, Scoring Formulas
Peer reviewedBurton, Richard F.; Miller, David J. – Assessment & Evaluation in Higher Education, 1999
Discusses statistical procedures for increasing test unreliability due to guessing in multiple choice and true/false tests. Proposes two new measures of test unreliability: one concerned with resolution of defined levels of knowledge and the other with the probability of examinees being incorrectly ranked. Both models are based on the binomial…
Descriptors: Guessing (Tests), Higher Education, Multiple Choice Tests, Objective Tests
Peer reviewedCizek, Gregory J.; Robinson, K. Lynne; O'Day, Denis M. – Educational and Psychological Measurement, 1998
The effect of removing nonfunctioning items from multiple-choice tests was studied by examining change in difficulty, discrimination, and dimensionality. Results provide additional support for the benefits of eliminating nonfunctioning options, such as enhanced score reliability, reduced testing time, potential for broader domain sampling, and…
Descriptors: Difficulty Level, Multiple Choice Tests, Sampling, Scores
Peer reviewedBurton, Richard F. – Assessment & Evaluation in Higher Education, 2001
Describes four measures of test unreliability that quantify effects of question selection and guessing, both separately and together--three chosen for immediacy and one for greater mathematical elegance. Quantifies their dependence on test length and number of answer options per question. Concludes that many multiple choice tests are unreliable…
Descriptors: Guessing (Tests), Mathematical Models, Multiple Choice Tests, Objective Tests
Peer reviewedTaylor, Annette Kujawski – College Student Journal, 2005
This research examined 2 elements of multiple-choice test construction, balancing the key and optimal number of options. In Experiment 1 the 3 conditions included a balanced key, overrepresentation of a and b responses, and overrepresentation of c and d responses. The results showed that error-patterns were independent of the key, reflecting…
Descriptors: Comparative Analysis, Test Items, Multiple Choice Tests, Test Construction
Zwick, Rebecca; And Others – 1993
Although the belief has been expressed that performance assessments are intrinsically more fair than multiple-choice measures, some forms of performance assessment may in fact be more likely than conventional tests to tap construct-irrelevant factors. As performance assessment grows in popularity, it will be increasingly important to monitor the…
Descriptors: Educational Assessment, Item Bias, Multiple Choice Tests, Performance Based Assessment
Lukhele, Robert; Sireci, Stephen G. – 1995
Free-response (FR) item formats, such as essay questions, are popular in educational assessment. The criticisms against FR items are that they are more expensive to score, take up more testing time, provide less content coverage, and are less reliable than multiple-choice (MC) items. For these reasons, FR items are often combined with MC items.…
Descriptors: Educational Assessment, Essay Tests, Item Response Theory, Multiple Choice Tests


