Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 2 |
| Since 2017 (last 10 years) | 4 |
| Since 2007 (last 20 years) | 6 |
Descriptor
| Error of Measurement | 20 |
| Multiple Choice Tests | 20 |
| Test Reliability | 20 |
| Higher Education | 6 |
| Scores | 5 |
| Test Items | 5 |
| Guessing (Tests) | 4 |
| Criterion Referenced Tests | 3 |
| Difficulty Level | 3 |
| Essay Tests | 3 |
| Item Response Theory | 3 |
| More ▼ | |
Source
Author
Publication Type
| Reports - Research | 11 |
| Journal Articles | 10 |
| Speeches/Meeting Papers | 4 |
| Reports - Evaluative | 3 |
| Reports - Descriptive | 2 |
| Dissertations/Theses -… | 1 |
| Guides - Non-Classroom | 1 |
| Numerical/Quantitative Data | 1 |
Education Level
| Grade 7 | 1 |
| High Schools | 1 |
| Higher Education | 1 |
| Middle Schools | 1 |
| Postsecondary Education | 1 |
Audience
| Researchers | 1 |
Location
| Ireland | 1 |
| United Kingdom (Wales) | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| General Educational… | 1 |
| New Jersey College Basic… | 1 |
| Test of Standard Written… | 1 |
What Works Clearinghouse Rating
Yu Wang – ProQuest LLC, 2024
The multiple-choice (MC) item format has been widely used in educational assessments across diverse content domains. MC items purportedly allow for collecting richer diagnostic information. The effectiveness and economy of administering MC items may have further contributed to their popularity not just in educational assessment. The MC item format…
Descriptors: Multiple Choice Tests, Cognitive Tests, Cognitive Measurement, Educational Diagnosis
Yousuf, Mustafa S.; Miles, Katherine; Harvey, Heather; Al-Tamimi, Mohammad; Badran, Darwish – Journal of University Teaching and Learning Practice, 2022
Exams should be valid, reliable, and discriminative. Multiple informative methods are used for exam analysis. Displaying analysis results numerically, however, may not be easily comprehended. Using graphical analysis tools could be better for the perception of analysis results. Two such methods were employed: standardized x-bar control charts with…
Descriptors: Multiple Choice Tests, Testing, Test Reliability, Test Validity
Al-zboon, Habis Saad; Alrekebat, Amjad Farhan – International Journal of Higher Education, 2021
This study aims at identifying the effect of multiple-choice test items' difficulty degree on the reliability coefficient and the standard error of measurement depending on the item response theory IRT. To achieve the objectives of the study, (WinGen3) software was used to generate the IRT parameters (difficulty, discrimination, guessing) for four…
Descriptors: Multiple Choice Tests, Test Items, Difficulty Level, Error of Measurement
Sheng, Yanyan – Measurement: Interdisciplinary Research and Perspectives, 2019
Classical approach to test theory has been the foundation for educational and psychological measurement for over 90 years. This approach concerns with measurement error and hence test reliability, which in part relies on individual test items. The CTT package, developed in light of this, provides functions for test- and item-level analyses of…
Descriptors: Item Response Theory, Test Reliability, Item Analysis, Error of Measurement
Irvin, P. Shawn; Alonzo, Julie; Lai, Cheng-Fei; Park, Bitnara Jasmine; Tindal, Gerald – Behavioral Research and Teaching, 2012
In this technical report, we present the results of a reliability study of the seventh-grade multiple choice reading comprehension measures available on the easyCBM learning system conducted in the spring of 2011. Analyses include split-half reliability, alternate form reliability, person and item reliability as derived from Rasch analysis,…
Descriptors: Reading Comprehension, Testing Programs, Statistical Analysis, Grade 7
Setzer, J. Carl – GED Testing Service, 2009
The GED[R] English as a Second Language (GED ESL) Test was designed to serve as an adjunct to the GED test battery when an examinee takes either the Spanish- or French-language version of the tests. The GED ESL Test is a criterion-referenced, multiple-choice instrument that assesses the functional, English reading skills of adults whose first…
Descriptors: Language Tests, High School Equivalency Programs, Psychometrics, Reading Skills
Peer reviewedCureton, Edward E. – Educational and Psychological Measurement, 1971
A rebuttal of Frary's 1969 article in Educational and Psychological Measurement. (MS)
Descriptors: Error of Measurement, Guessing (Tests), Multiple Choice Tests, Scoring Formulas
Peer reviewedZimmerman, Donald W. – Educational and Psychological Measurement, 1985
A computer program simulated guessing on multiple-choice test items and calculated deviation IQ's from observed scores which contained a guessing component. Extensive variability in deviation IQ's due entirely to chance was found. (Author/LMO)
Descriptors: Computer Simulation, Error of Measurement, Guessing (Tests), Intelligence Quotient
Peer reviewedZimmerman, Donald W.; And Others – Journal of Experimental Education, 1984
Three types of test were compared: a completion test, a matching test, and a multiple-choice test. The completion test was more reliable than the matching test, and the matching test was more reliable than the multiple-choice test. (Author/BW)
Descriptors: Comparative Analysis, Error of Measurement, Higher Education, Mathematical Models
Frary, Robert B.; Zimmerman, Donald W. – Educ Psychol Meas, 1970
Descriptors: Error of Measurement, Guessing (Tests), Multiple Choice Tests, Probability
Peer reviewedStraton, Ralph G.; Catts, Ralph M. – Educational and Psychological Measurement, 1980
Multiple-choice tests composed entirely of two-, three-, or four-choice items were investigated. Results indicated that number of alternatives per item was inversely related to item difficulty, but directly related to item discrimination. Reliability and standard error of measurement of three-choice item tests was equivalent or superior.…
Descriptors: Difficulty Level, Error of Measurement, Foreign Countries, Higher Education
Peer reviewedWerts, C. E.; And Others – Educational and Psychological Measurement, 1980
Test-retest correlations can lead to biased reliability estimates when there is instability of true scores and/or when measurement errors are correlated. Using three administrations of the Test of Standard Written English and essay ratings, an analysis is demonstrated which separates true score instability and correlated errors. (Author/BW)
Descriptors: College Freshmen, Error of Measurement, Essay Tests, Higher Education
Schmitt, Alicia P.; Crocker, Linda – 1981
The effectiveness of a strategy for improving performance on multiple choice items for examinees with different levels of test anxiety was assessed. Undergraduate measurement students responded to the Mandler-Sarason Test Anxiety Scale and to an objective test covering course content. Results indicated that, for most examinees, generation of an…
Descriptors: Error of Measurement, Higher Education, Multiple Choice Tests, Response Style (Tests)
Scheetz, James P.; vonFraunhofer, J. Anthony – 1980
Subkoviak suggested a technique for estimating both group reliability and the reliability associated with assigning a given individual to a mastery or non-mastery category based on a single test administration. Two assumptions underlie this model. First, it is assumed that had successive test administrations occurred, scores for each individual…
Descriptors: Criterion Referenced Tests, Cutting Scores, Error of Measurement, Higher Education
Sykes, Robert C.; Hou, Liling – Applied Measurement in Education, 2003
Weighting responses to Constructed-Response (CR) items has been proposed as a way to increase the contribution these items make to the test score when there is insufficient testing time to administer additional CR items. The effect of various types of weighting items of an IRT-based mixed-format writing examination was investigated.…
Descriptors: Item Response Theory, Weighted Scores, Responses, Scores
Previous Page | Next Page ยป
Pages: 1 | 2
Direct link
