Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 0 |
| Since 2007 (last 20 years) | 1 |
Descriptor
| Scoring Formulas | 7 |
| Statistical Studies | 7 |
| Difficulty Level | 4 |
| Item Analysis | 4 |
| Mathematical Models | 3 |
| Test Items | 3 |
| Academic Ability | 2 |
| Error of Measurement | 2 |
| Guessing (Tests) | 2 |
| High Schools | 2 |
| Higher Education | 2 |
| More ▼ | |
Author
| Ben-Simon, Anat | 1 |
| Bennett, Randy Elliott | 1 |
| Donlon, Thomas F. | 1 |
| Fitzpatrick, Anne R. | 1 |
| Hilton, Thomas L. | 1 |
| Kingston, Neal M. | 1 |
| Livingston, Samuel A. | 1 |
| Lord, Frederic M. | 1 |
| Mitchell, Virginia P. | 1 |
| Smith, Richard M. | 1 |
Publication Type
| Reports - Research | 7 |
| Speeches/Meeting Papers | 5 |
| Journal Articles | 2 |
Education Level
| Elementary Education | 1 |
| Grade 8 | 1 |
Audience
| Researchers | 4 |
Location
Laws, Policies, & Programs
Assessments and Surveys
| Graduate Record Examinations | 1 |
| National Longitudinal Study… | 1 |
What Works Clearinghouse Rating
Peer reviewedLord, Frederic M. – Journal of Educational Measurement, 1984
Four methods are outlined for estimating or approximating from a single test administration the standard error of measurement of number-right test score at specified ability levels or cutting scores. The methods are illustrated and compared on one set of real test data. (Author)
Descriptors: Academic Ability, Cutting Scores, Error of Measurement, Scoring Formulas
Ben-Simon, Anat; Bennett, Randy Elliott – Journal of Technology, Learning, and Assessment, 2007
This study evaluated a "substantively driven" method for scoring NAEP writing assessments automatically. The study used variations of an existing commercial program, e-rater[R], to compare the performance of three approaches to automated essay scoring: a "brute-empirical" approach in which variables are selected and weighted solely according to…
Descriptors: Writing Evaluation, Writing Tests, Scoring, Essays
Livingston, Samuel A. – 1986
This paper deals with test fairness regarding a test consisting of two parts: (1) a "common" section, taken by all students; and (2) a "variable" section, in which some students may answer a different set of questions from other students. For example, a test taken by several thousand students each year contains a common multiple-choice portion and…
Descriptors: Difficulty Level, Error of Measurement, Essay Tests, Mathematical Models
Donlon, Thomas F.; Fitzpatrick, Anne R. – 1978
On the basis of past research efforts to improve multiple-choice test information through differential weighting of responses to wrong answers (distractors), two statistical indices are developed. Each describes the properties of response distributions across the options of an item. Jaspen's polyserial generalization of the biserial correlation…
Descriptors: Confidence Testing, Difficulty Level, Guessing (Tests), High Schools
Kingston, Neal M. – 1985
Birnbaum's three-parameter logistic item response model was used to study guessing behavior of low ability examinees on the Graduate Record Examinations (GRE) General Test, Verbal Measure. GRE scoring procedures had recently changed, from a scoring formula which corrected for guessing, to number-right scoring. The three-parameter theory was used…
Descriptors: Academic Aptitude, Analysis of Variance, College Entrance Examinations, Difficulty Level
Smith, Richard M.; Mitchell, Virginia P. – 1979
To improve the accuracy of college placement, Rasch scoring and person-fit statistics on the Comparative Guidance and Placement test (CGP) was compared to the traditional right-only scoring. Correlations were calculated between English and mathematics course grades and scores of 1,448 entering freshmen on the reading, writing, and mathematics…
Descriptors: Academic Ability, Computer Programs, Difficulty Level, Goodness of Fit
Hilton, Thomas L.; And Others – 1985
Since the mean score for a sample composed of several subgroups can be viewed as the sum of the mean of each subgroup weighted by the proportional size of the subgroup, then the mean change in a time period--in this case, from 1972 to 1980--is the sum of the differences between the means for each subgroup, with each mean weighted by its…
Descriptors: Analysis of Covariance, Cohort Analysis, Cross Sectional Studies, Educational Trends

Direct link
