Publication Date
| In 2026 | 0 |
| Since 2025 | 10 |
| Since 2022 (last 5 years) | 40 |
| Since 2017 (last 10 years) | 118 |
| Since 2007 (last 20 years) | 211 |
Descriptor
| Multiple Choice Tests | 532 |
| Test Reliability | 532 |
| Test Validity | 302 |
| Test Construction | 238 |
| Test Items | 172 |
| Foreign Countries | 114 |
| Item Analysis | 101 |
| Higher Education | 90 |
| Difficulty Level | 85 |
| Guessing (Tests) | 74 |
| Scoring | 69 |
| More ▼ | |
Source
Author
| Ebel, Robert L. | 10 |
| Frary, Robert B. | 9 |
| Alonzo, Julie | 7 |
| Frisbie, David A. | 6 |
| Irvin, P. Shawn | 6 |
| Lai, Cheng-Fei | 6 |
| Park, Bitnara Jasmine | 6 |
| Tindal, Gerald | 6 |
| Wilcox, Rand R. | 5 |
| Albanese, Mark A. | 4 |
| Biancarosa, Gina | 4 |
| More ▼ | |
Publication Type
Education Level
Audience
| Researchers | 11 |
| Practitioners | 8 |
| Teachers | 5 |
Location
| Indonesia | 17 |
| Turkey | 17 |
| Germany | 8 |
| Iran | 8 |
| Canada | 6 |
| Malaysia | 4 |
| Nigeria | 4 |
| Australia | 3 |
| Florida | 3 |
| Japan | 3 |
| Pakistan | 3 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Peer reviewedGibbons, Jean D.; And Others – Psychometrika, 1979
On a multiple-choice test in which each item has k alternative responses, the test taker is permitted to choose any subset which he believes contains the one correct answer. A scoring system is devised. (Author/CTM)
Descriptors: Confidence Testing, Efficiency, Multiple Choice Tests, Scoring
Peer reviewedNishisato, Shizuhiko; Sheu, Wen-Jenn – Psychometrika, 1980
A modification of the method of reciprocal averages for scaling multiple choice data is proposed. The proposed method handles the data in a piecewise fashion and allows for faster convergence to a solution. (Author/JKS)
Descriptors: Item Analysis, Measurement Techniques, Multiple Choice Tests, Test Reliability
Trieber, J. Marshall – Training and Development Journal, 1980
Aims to help instructors make more valid test questions, particularly multiple-choice ones. Emphasis is placed on multiple-choice questions to show the wealth of opportunities they offer for testing because of their uses, objectivity, and ease of grading. Discusses test scheduling, construction, and evaluation and follow-up. (CT)
Descriptors: Multiple Choice Tests, Test Construction, Test Reliability, Test Validity
Peer reviewedHolmes, Roy A.; And Others – Educational and Psychological Measurement, 1974
Descriptors: Chemistry, Multiple Choice Tests, Scoring Formulas, Test Reliability
Strasler, Gregg M.; Raeth, Peter G. – 1977
The study investigated the feasibility of adapting the coefficient k introduced by Cohen (1960) and elaborated by Swaminathan, Hambleton, and Algina (1974) to an internal consistency estimate for criterion referenced tests in single test administrations. The authors proposed the use of k as an internal consistency estimate by logically dividing…
Descriptors: Computer Programs, Criterion Referenced Tests, Multiple Choice Tests, Test Reliability
Peer reviewedGilman, David Alan; Ferry, Paula – Journal of Educational Measurement, 1972
Results indicate that scoring tests by the self-scoring method can result in a higher split half reliability than tests scored by the traditional right-wrong method. (Authors)
Descriptors: Data Analysis, Multiple Choice Tests, Scoring, Test Construction
Sabers, Darrell L.; White, Gordon W. – J Educ Meas, 1969
Descriptors: Aptitude Tests, Multiple Choice Tests, Predictive Validity, Scoring
Ebel, Robert L. – Educ Psychol Meas, 1969
Descriptors: Item Analysis, Multiple Choice Tests, Objective Tests, Test Reliability
Rich, Charles E.; Johanson, George A. – 1990
Despite the existence of little empirical evidence for their effectiveness, many techniques have been suggested for writing multiple-choice items. The option "none of the above" (NA) has been widely used although a recent review of empirical studies of NA suggests that, while generally decreasing the difficulty index, NA also decreases…
Descriptors: Difficulty Level, Item Analysis, Multiple Choice Tests, Test Construction
Incremental Reliability and Validity of Multiple-Choice Tests with an Answer-Until-Correct Procedure
Peer reviewedHanna, Gerald S. – Journal of Educational Measurement, 1975
An alternative to the conventional right-wrong scoring method used on multiple-choice tests was presented. In the experiment, the examinee continued to respond to a multiple-choice item until feedback signified a correct answer. Findings showed that experimental scores were more reliable but less valid than inferred conventional scores.…
Descriptors: Feedback, Higher Education, Multiple Choice Tests, Scoring
Peer reviewedRaffeld, Paul – Journal of Educational Measurement, 1975
Results support the contention that a Guttman-weighted objective test can have psychometric properties that are superior to those of its unweighted counterpart, as long as omissions do not exist or are assigned a value equal to the mean of the k item alternative weights. (Author/BJG)
Descriptors: Multiple Choice Tests, Predictive Validity, Test Reliability, Test Validity
Peer reviewedCrehan, Kevin D.; Slakter, Malcolm J. – Psychological Reports, 1971
Descriptors: Comparative Analysis, Multiple Choice Tests, Test Construction, Test Reliability
Hambleton, Ronald K.; And Others – J Educ Meas, 1970
Descriptors: Comparative Analysis, Evaluation Methods, Multiple Choice Tests, Test Reliability
Peer reviewedWilcox, Rand R. – Educational and Psychological Measurement, 1982
Results in the engineering literature on "k out of n system reliability" can be used to characterize tests based on estimates of the probability of correctly determining whether the examinee knows the correct response. In particular, the minimum number of distractors required for multiple-choice tests can be empirically determined.…
Descriptors: Achievement Tests, Mathematical Models, Multiple Choice Tests, Test Format
Attali, Yigal – ETS Research Report Series, 2004
Contrary to common belief, reliability estimates of number-right multiple-choice tests are not inflated by speededness. Because examinees guess on questions when they run out of time, the responses to these questions show less consistency with the responses of other questions, and the reliability of the test will be decreased. The surprising…
Descriptors: Multiple Choice Tests, Timed Tests, Test Reliability, Guessing (Tests)


