Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 2 |
| Since 2017 (last 10 years) | 11 |
| Since 2007 (last 20 years) | 24 |
Descriptor
| Scoring Formulas | 146 |
| Test Reliability | 146 |
| Test Validity | 66 |
| Multiple Choice Tests | 47 |
| Guessing (Tests) | 38 |
| Test Construction | 33 |
| Test Interpretation | 26 |
| Test Items | 25 |
| Higher Education | 23 |
| Scoring | 23 |
| Item Analysis | 22 |
| More ▼ | |
Source
Author
Publication Type
Education Level
| Higher Education | 7 |
| Postsecondary Education | 6 |
| Elementary Education | 2 |
| Elementary Secondary Education | 2 |
| Secondary Education | 2 |
| Adult Education | 1 |
| High Schools | 1 |
| Junior High Schools | 1 |
| Middle Schools | 1 |
Audience
| Researchers | 2 |
| Practitioners | 1 |
Location
| New York (New York) | 2 |
| Australia | 1 |
| Canada | 1 |
| Germany | 1 |
| India | 1 |
| Malaysia | 1 |
| Minnesota | 1 |
| Mississippi | 1 |
| New York | 1 |
| North Carolina | 1 |
| Ohio | 1 |
| More ▼ | |
Laws, Policies, & Programs
| Elementary and Secondary… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Peer reviewedKane, Michael; Moloney, James – Applied Psychological Measurement, 1978
The answer-until-correct (AUC) procedure requires that examinees respond to a multi-choice item until they answer it correctly. Using a modified version of Horst's model for examinee behavior, this paper compares the effect of guessing on item reliability for the AUC procedure and the zero-one scoring procedure. (Author/CTM)
Descriptors: Guessing (Tests), Item Analysis, Mathematical Models, Multiple Choice Tests
Peer reviewedEakin, Richard R.; Long, Clifford A. – Educational and Psychological Measurement, 1977
A scoring technique for true-false tests is presented. The technique, paired item scoring, involves combining two statements and having the student select one of the four resultants possible: true-true, false-true, true-false, and false-false. The combined item is treated as a multiple choice item. (Author/JKS)
Descriptors: Guessing (Tests), Measurement Techniques, Multiple Choice Tests, Objective Tests
Peer reviewedHuynh, Huynh – Journal of Educational Statistics, 1986
Under the assumptions of classical measurement theory and the condition of normality, a formula is derived for the reliability of composite scores. The formula represents an extension of the Spearman-Brown formula to the case of truncated data. (Author/JAZ)
Descriptors: Computer Simulation, Error of Measurement, Expectancy Tables, Scoring Formulas
Peer reviewedHocevar, Dennis; Michael, William B. – Educational and Psychological Measurement, 1979
Two multitrait-multimethod studies were conducted to investigate the effects of two scoring formulas. The study demonstrates that tests of divergent thinking lack discriminant validity when scored in the usual manner. A percentage formula did enhance discriminant validity when originality ratings were subjectively determined. (Author/CTM)
Descriptors: Creativity, Creativity Tests, Divergent Thinking, Grade 5
Knapp, Thomas R. – Measurement and Evaluation in Guidance, 1980
Supports arguments against general use of change scores and recommends the Lord/McNemar estimates of true change. Provides a numerical example illustrating the reliability problem and the problem of the prediction of true change from various linear composites of initial and final measures. (Author)
Descriptors: Counseling Techniques, Literature Reviews, Pretests Posttests, Research Methodology
Peer reviewedStauffer, A. J. – Educational and Psychological Measurement, 1974
Descriptors: Attitude Change, Attitude Measures, Comparative Analysis, Educational Research
Halpin, Gerald; Simpson, Robert – 1986
Forty adolescent subjects with behavior problems at home or at school were administered the Woodcock Reading Mastery Tests. Subjects ranged in age from 12 to 18, and included 25 males and 15 females. The Passage Comprehension Test for each subject was rescored using three different ceiling criteria: (1) five errors in six consecutive responses,…
Descriptors: Adolescents, Analysis of Variance, Behavior Problems, Raw Scores
Peer reviewedEchternacht, Gary – Educational and Psychological Measurement, 1976
Compares various item option scoring methods with respect to coefficient alpha and a concurrent validity coefficient. Scoring methods compared were: formula scoring, a priori scoring, empirical scoring with an internal criterion, and two modifications of formula scoring. The empirically determined scoring system is seen as superior. (RC)
Descriptors: Aptitude Tests, Multiple Choice Tests, Response Style (Tests), Scoring Formulas
Peer reviewedTraub, Ross E.; Hambleton, Ronald K. – Educational and Psychological Measurement, 1973
Descriptors: Grade 8, Guessing (Tests), Multiple Choice Tests, Pacing
Peer reviewedJacobs, Stanley S. – Journal of Educational Measurement, 1971
Descriptors: Guessing (Tests), Individual Differences, Measurement Techniques, Multiple Choice Tests
Peer reviewedAlbanese, Mark A. – Evaluation and the Health Professions, 1982
Findings regarding formats and scoring formulas for multiple-choice test items with more than one correct response are presented. Strong cluing effects in the Type K format, increasing the correct score percentage and reducing test reliability, recommend using the Type X format. Alternative scoring methods are discussed. (Author/CM)
Descriptors: Health Occupations, Multiple Choice Tests, Professional Education, Response Style (Tests)
Peer reviewedAiken, Lewis R. – Educational and Psychological Measurement, 1980
Procedures for computing content validity and consistency reliability coefficients and determining the statistical significance of these coefficients are described. Procedures employing the multinomial probability distribution for small samples and normal curve probability estimates for large samples, can be used where judgments are made on…
Descriptors: Computer Programs, Measurement Techniques, Probability, Questionnaires
Peer reviewedDiamond, James J. – Journal of Educational Measurement, 1975
Investigates the reliability and validity of scores yielded from a new scoring formula. (Author/DEP)
Descriptors: Guessing (Tests), Multiple Choice Tests, Objective Tests, Scoring
Hendrickson, Gerry F.; Green, Bert F., Jr. – 1972
It has been shown that Guttman weighting of test options results in marked increases in the internal consistency of a test. However, the effect of this type of weighting on the structure of the test is not known. Hence, the purpose of this study is to compare the factor structure of Guttman-weighted and rights-only-weighted tests and to relate the…
Descriptors: Analysis of Variance, Correlation, Factor Analysis, Item Analysis
Peer reviewedPoizner, Sharon B.; And Others – Applied Psychological Measurement, 1978
Binary, probability, and ordinal scoring procedures for multiple-choice items were examined. In two situations, it was found that both the probability and ordinal scoring systems were more reliable than the binary scoring method. (Author/CTM)
Descriptors: Confidence Testing, Guessing (Tests), Higher Education, Multiple Choice Tests


