Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 0 |
| Since 2007 (last 20 years) | 2 |
Descriptor
| Comparative Testing | 3 |
| Standard Setting (Scoring) | 3 |
| Cutting Scores | 2 |
| Equated Scores | 2 |
| Item Analysis | 2 |
| Construct Validity | 1 |
| Estimation (Mathematics) | 1 |
| Evaluators | 1 |
| Grade 8 | 1 |
| Investigations | 1 |
| Licensing Examinations… | 1 |
| More ▼ | |
Author
| Ferdous, Abdullah A. | 1 |
| Koskey, Kristin L. K. | 1 |
| Norcini, John | 1 |
| Plake, Barbara S. | 1 |
| Sondergeld, Toni A. | 1 |
| Stone, Gregory Ethan | 1 |
Publication Type
| Journal Articles | 3 |
| Reports - Evaluative | 3 |
Education Level
| Elementary Education | 1 |
| Elementary Secondary Education | 1 |
| Grade 8 | 1 |
| Higher Education | 1 |
Audience
Location
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Stone, Gregory Ethan; Koskey, Kristin L. K.; Sondergeld, Toni A. – Educational and Psychological Measurement, 2011
Typical validation studies on standard setting models, most notably the Angoff and modified Angoff models, have ignored construct development, a critical aspect associated with all conceptualizations of measurement processes. Stone compared the Angoff and objective standard setting (OSS) models and found that Angoff failed to define a legitimate…
Descriptors: Cutting Scores, Standard Setting (Scoring), Models, Construct Validity
Ferdous, Abdullah A.; Plake, Barbara S. – Educational and Psychological Measurement, 2007
In an Angoff standard setting procedure, judges estimate the probability that a hypothetical randomly selected minimally competent candidate will answer correctly each item in the test. In many cases, these item performance estimates are made twice, with information shared with the panelists between estimates. Especially for long tests, this…
Descriptors: Test Items, Probability, Item Analysis, Standard Setting (Scoring)
Peer reviewedNorcini, John; And Others – Applied Psychological Measurement, 1991
Effects of numbers of experts (NOEs) and common items (CIs) on the scaling of cutting scores from expert judgments were studied for 11,917 physicians taking 2 forms of a medical specialty examination. Increasing NOEs and CIs reduced error; beyond 5 experts and 25 CIs, error differences were small. (SLD)
Descriptors: Comparative Testing, Cutting Scores, Equated Scores, Estimation (Mathematics)

Direct link
