Publication Date
| In 2026 | 0 |
| Since 2025 | 85 |
| Since 2022 (last 5 years) | 453 |
| Since 2017 (last 10 years) | 1241 |
| Since 2007 (last 20 years) | 2515 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 122 |
| Teachers | 105 |
| Researchers | 64 |
| Students | 46 |
| Administrators | 14 |
| Policymakers | 7 |
| Counselors | 3 |
| Parents | 3 |
Location
| Canada | 134 |
| Turkey | 131 |
| Australia | 123 |
| Iran | 66 |
| Indonesia | 61 |
| United Kingdom | 51 |
| Germany | 50 |
| Taiwan | 46 |
| United States | 43 |
| China | 39 |
| California | 35 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 3 |
| Meets WWC Standards with or without Reservations | 5 |
| Does not meet standards | 6 |
Peer reviewedAlbanese, Mark A. – Evaluation and the Health Professions, 1982
Findings regarding formats and scoring formulas for multiple-choice test items with more than one correct response are presented. Strong cluing effects in the Type K format, increasing the correct score percentage and reducing test reliability, recommend using the Type X format. Alternative scoring methods are discussed. (Author/CM)
Descriptors: Health Occupations, Multiple Choice Tests, Professional Education, Response Style (Tests)
Peer reviewedGross, Leon J. – Journal of Optometric Education, 1982
A critique of a variety of formats used in combined-response test items (those in which the respondent must choose the correct combination of options: a and b, all of the above, etc.) illustrates why this kind of testing is inherently flawed and should not be used in optometry examinations. (MSE)
Descriptors: Higher Education, Multiple Choice Tests, Optometry, Standardized Tests
Peer reviewedBodner, George M. – Journal of Chemical Education, 1980
Presented are common words and phrases which are encountered in the statistical analysis of test results. Included are analysis of the mid-point, distribution of scores, calculation of scaled scores, test reliability, item analysis, and coefficients of reliability. (CS)
Descriptors: College Science, Definitions, Evaluation Methods, Higher Education
Peer reviewedWhetton, C.; Childs, R. – British Journal of Educational Psychology, 1981
Answer-until-correct (AUC) is a procedure for providing feedback during a multiple-choice test, giving an increased range of scores. The performance of secondary students on a verbal ability test using AUC procedures was compared with a group using conventional instructions. AUC scores considerably enhanced reliability but not validity.…
Descriptors: Feedback, Multiple Choice Tests, Response Style (Tests), Secondary Education
Peer reviewedIrvin, Larry K.; And Others – Journal of Educational Measurement, 1980
The relative efficacy of content-appropriate, orally administered true/false and multiple-choice testing was examined with retarded adolescents. Both approaches demonstrated utility and psychometric adequacy. Implications regarding test development for retarded students are briefly discussed. (Author)
Descriptors: High Schools, Mild Mental Retardation, Multiple Choice Tests, Objective Tests
Peer reviewedStraton, Ralph G.; Catts, Ralph M. – Educational and Psychological Measurement, 1980
Multiple-choice tests composed entirely of two-, three-, or four-choice items were investigated. Results indicated that number of alternatives per item was inversely related to item difficulty, but directly related to item discrimination. Reliability and standard error of measurement of three-choice item tests was equivalent or superior.…
Descriptors: Difficulty Level, Error of Measurement, Foreign Countries, Higher Education
Peer reviewedWerts, C. E.; And Others – Educational and Psychological Measurement, 1980
Test-retest correlations can lead to biased reliability estimates when there is instability of true scores and/or when measurement errors are correlated. Using three administrations of the Test of Standard Written English and essay ratings, an analysis is demonstrated which separates true score instability and correlated errors. (Author/BW)
Descriptors: College Freshmen, Error of Measurement, Essay Tests, Higher Education
Peer reviewedGreen, Kathy – Journal of Experimental Education, 1979
Reliabilities and concurrent validities of teacher-made multiple-choice and true-false tests were compared. No significant differences were found even when multiple-choice reliability was adjusted to equate testing time. (Author/MH)
Descriptors: Comparative Testing, Higher Education, Multiple Choice Tests, Test Format
Peer reviewedAnderson, Johnston – Educational Studies in Mathematics, 1979
An account is given of the introduction of multiple-choice and other forms of objective testing into the assessment of a first-year university elementary analysis class. (MP)
Descriptors: Achievement, Calculus, College Mathematics, Evaluation Methods
Peer reviewedLiskin-Gasparro, Judith; And Others – Hispania, 1979
Reports on a study which sought to determine the validity of the Advanced Placement Spanish Language Examination by comparing the performance on the multiple-choice section of this test of three groups: (1) advanced placement students, (2) second-year college students, and (3) third-year college students. (AM)
Descriptors: Advanced Placement, Higher Education, Language Tests, Multiple Choice Tests
Peer reviewedWeitzman, R. A. – Educational and Psychological Measurement, 1996
Application of the Rasch model to data from 267 naval personnel taking a multiple-choice test subject to guessing indicates that goodness of fit is a problem small enough to ignore, but that modification to account for guessing may be necessary for accurate equating. (SLD)
Descriptors: Equated Scores, Goodness of Fit, Guessing (Tests), Item Response Theory
Peer reviewedPajares, Frank; Miller, M. David – Journal of Experimental Education, 1997
The mathematics self-efficacy and problem-solving performance of 327 middle school students were assessed with multiple-choice and open-ended methods. No differences in self-efficacy resulted from the different forms of assessment, although those who took the multiple-choice test had higher scores and better calibration of ability. (SLD)
Descriptors: Ability, Educational Assessment, Mathematics, Middle School Students
Peer reviewedHerman, William E. – Journal of Research and Development in Education, 1997
This study explored the relationship between time needed to complete multiple choice tests and college student test performance. Examination time and performance for midsemester and final exams were measured. Results indicated that the variables were relatively unrelated. Students tended to use a consistent test-taking tempo on relatively untimed…
Descriptors: Academic Achievement, College Students, Higher Education, Multiple Choice Tests
Peer reviewedBeullens, Johan; Van Damme, Bo; Jaspaert, Hendrik; Janssen, Piet J. – Medical Teacher, 2002
Explains the evaluation procedure of extended matching multiple-choice questions (EMQs) which are problem-oriented, case-based questions. (Author/YDS)
Descriptors: Case Method (Teaching Technique), Evaluation Methods, Higher Education, Medical Education
Peer reviewedJaradat, Derar; Tollefson, Nona – Educational and Psychological Measurement, 1988
This study compared the reliability and validity indexes of randomly parallel tests administered under inclusion, exclusion, and correction for guessing directions, using 54 graduate students. It also compared the criterion-referenced grading decisions based on the different scoring methods. (TJH)
Descriptors: Criterion Referenced Tests, Grading, Graduate Students, Guessing (Tests)


