Publication Date
| In 2026 | 0 |
| Since 2025 | 1 |
| Since 2022 (last 5 years) | 1 |
| Since 2017 (last 10 years) | 4 |
| Since 2007 (last 20 years) | 4 |
Descriptor
| Multiple Choice Tests | 41 |
| Test Reliability | 41 |
| Testing Problems | 41 |
| Test Validity | 22 |
| Test Construction | 18 |
| Guessing (Tests) | 13 |
| Test Items | 11 |
| Response Style (Tests) | 10 |
| Item Analysis | 8 |
| Objective Tests | 8 |
| Scoring | 8 |
| More ▼ | |
Source
Author
| Ebel, Robert L. | 3 |
| Wilcox, Rand R. | 2 |
| Anderson, Paul S. | 1 |
| Bao, Lei | 1 |
| Bracey, Gerald W. | 1 |
| Brady, Raymond G. | 1 |
| Carruthers, Robert B. | 1 |
| Cesur, Kursat | 1 |
| Cizek, Gregory J. | 1 |
| Coats, Pamela K. | 1 |
| Crocker, Linda | 1 |
| More ▼ | |
Publication Type
Education Level
Audience
| Practitioners | 1 |
| Researchers | 1 |
Location
| California | 1 |
| Colombia | 1 |
| Indonesia | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| National Assessment of… | 1 |
What Works Clearinghouse Rating
Cesur, Kursat – Educational Policy Analysis and Strategic Research, 2019
Examinees' performances are assessed using a wide variety of different techniques. Multiple-choice (MC) tests are among the most frequently used ones. Nearly, all standardized achievement tests make use of MC test items and there is a variety of ways to score these tests. The study compares number right and liberal scoring (SAC) methods. Mixed…
Descriptors: Multiple Choice Tests, Scoring, Evaluation Methods, Guessing (Tests)
Patrisius Istiarto Djiwandono; Daniel Ginting – Language Education & Assessment, 2025
The teaching of English as a foreign language in Indonesia has a long history, and it is always important to ask whether the assessment of the students' language skills has been valid and reliable. A screening of many articles in several prominent databases reveal that a number of evaluation studies have been done by Indonesian scholars in the…
Descriptors: Foreign Countries, Language Tests, English (Second Language), Second Language Learning
Xiao, Yang; Han, Jing; Koenig, Kathleen; Xiong, Jianwen; Bao, Lei – Physical Review Physics Education Research, 2018
Assessment instruments composed of two-tier multiple choice (TTMC) items are widely used in science education as an effective method to evaluate students' sophisticated understanding. In practice, however, there are often concerns regarding the common scoring methods of TTMC items, which include pair scoring and individual scoring schemes. The…
Descriptors: Hierarchical Linear Modeling, Item Response Theory, Multiple Choice Tests, Case Studies
Giraldo, Frank – HOW, 2019
The purpose of this article of reflection is to raise awareness of how poor design of language assessments may have detrimental effects, if crucial qualities and technicalities of test design are not met. The article first discusses these central qualities for useful language assessments. Then, guidelines for creating listening assessments, as an…
Descriptors: Test Construction, Consciousness Raising, Language Tests, Second Language Learning
Garvin, Alfred D.; Ralston, Nancy C. – 1970
Confidence Weighting (CW), after Ebel, and Multiple Responding (MR), after Coombs, are compared empirically to determine which improved test reliability more in the case of a course pretest derived from the final examination. It was hypothesized that MR, which purportedly measures partial knowledge, would be more effective than CW, which measures…
Descriptors: College Students, Educational Testing, Multiple Choice Tests, Objective Tests
Peer reviewedLord, Frederic M. – Journal of Educational Measurement, 1977
Two approaches for determining the optimal number of choices for a test item, presently in the literature, are compared with two new approaches. (Author)
Descriptors: Forced Choice Technique, Latent Trait Theory, Multiple Choice Tests, Test Items
Cizek, Gregory J. – Phi Delta Kappan, 1991
This rejoinder to Grant Wiggins on performance assessment suggests that true educational reform will undoubtedly be evidenced by something more substantial than pocket folders bulging with student work. Labeling performance tests "authentic" does not ensure their validity, reliability, or incorruptibility. Such tests are neither replacements nor…
Descriptors: Elementary Secondary Education, Multiple Choice Tests, Performance Based Assessment, Pilot Projects
Miller, Harry G.; Williams, Reed G. – Educational Technology, 1973
Descriptors: Content Analysis, Item Analysis, Measurement Techniques, Multiple Choice Tests
Schmitt, Alicia P.; Crocker, Linda – 1981
The effectiveness of a strategy for improving performance on multiple choice items for examinees with different levels of test anxiety was assessed. Undergraduate measurement students responded to the Mandler-Sarason Test Anxiety Scale and to an objective test covering course content. Results indicated that, for most examinees, generation of an…
Descriptors: Error of Measurement, Higher Education, Multiple Choice Tests, Response Style (Tests)
Ebel, Robert L. – 1973
True-false achievement test items written by typical classroom teachers show about two-thirds of the discrimination of their multiple-choice test items. This is about what should be expected in view of the higher probability of chance success on the true-false items. However, at least half again as many true-false items as multiple-choice items…
Descriptors: Guessing (Tests), Multiple Choice Tests, Objective Tests, Scoring
Peer reviewedWhitby, L. G. – Medical Education, 1977
Advantages and disadvantages of no-penalty and penalty marking systems are discussed. Ways in which examiners have attempted to correct for guessing by students are reviewed, along with the use of "don't know" options and confidence-weighting for attempting to assess the degree of certainty that candidates attach to their answers. (Author/LBH)
Descriptors: Grading, Guessing (Tests), Higher Education, Medical Education
Peer reviewedYelvington, James Yowell; Brady, Raymond G. – Community/Junior College Research Quarterly, 1979
Assesses the applicability of corrective feedback (CF) testing, which allows multiple attempts to respond to a test item, to the community college classroom. Compares CF testing to single answer testing, especially with regard to reliability, equitability, and effect on student motivation. (DD)
Descriptors: Community Colleges, Educational Testing, Feedback, Multiple Choice Tests
Koehler, Roger A. – 1974
A potentially valuable measure of overconfidence on probabilistic multiple-choice tests was evaluated. The measure of overconfidence was based on probabilistic responses to nonsense items embedded in a vocabulary test. The test was administered under both confidence response and conventional choice response directions to 208 undergraduate…
Descriptors: Confidence Testing, Guessing (Tests), Measurement Techniques, Multiple Choice Tests
Myers, Charles T. – 1978
The viewpoint is expressed that adding to test reliability by either selecting a more homogeneous set of items, restricting the range of item difficulty as closely as possible to the most efficient level, or increasing the number of items will not add to test validity and that there is considerable danger that efforts to increase reliability may…
Descriptors: Achievement Tests, Item Analysis, Multiple Choice Tests, Test Construction
Peer reviewedEbel, Robert L. – Educational and Psychological Measurement, 1978
A multiple true-false item is one where a testee has to identify statements as true or false within a cluster (of two or more) of such statements. Clusters are then scored as items. This study showed such a procedure to yield less reliable results than traditional true-false items. (JKS)
Descriptors: Guessing (Tests), Higher Education, Item Analysis, Multiple Choice Tests

Direct link
