Publication Date
| In 2026 | 0 |
| Since 2025 | 5 |
| Since 2022 (last 5 years) | 53 |
| Since 2017 (last 10 years) | 155 |
| Since 2007 (last 20 years) | 553 |
Descriptor
Source
Author
| Hambleton, Ronald K. | 37 |
| Popham, W. James | 30 |
| Ediger, Marlow | 15 |
| Roid, Gale | 12 |
| Baker, Eva L. | 11 |
| Wilcox, Rand R. | 11 |
| Berk, Ronald A. | 10 |
| Haladyna, Tom | 10 |
| Livingston, Samuel A. | 10 |
| Millman, Jason | 10 |
| Nitko, Anthony J. | 10 |
| More ▼ | |
Publication Type
Education Level
Audience
| Practitioners | 255 |
| Teachers | 141 |
| Researchers | 99 |
| Administrators | 31 |
| Policymakers | 17 |
| Parents | 13 |
| Community | 9 |
| Students | 6 |
| Counselors | 2 |
| Support Staff | 1 |
Location
| Georgia | 85 |
| Australia | 50 |
| Florida | 36 |
| Missouri | 31 |
| Texas | 28 |
| Canada | 24 |
| Oklahoma | 18 |
| Illinois | 17 |
| United States | 17 |
| South Carolina | 16 |
| California | 15 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Does not meet standards | 5 |
Coe, Robert – Oxford Review of Education, 2008
The comparability of examinations in different subjects has been a controversial topic for many years and a number of criticisms have been made of statistical approaches to estimating the "difficulties" of achieving particular grades in different subjects. This paper argues that if comparability is understood in terms of a linking…
Descriptors: Test Items, Grades (Scholastic), Foreign Countries, Test Bias
Hall, John D.; Howerton, D. Lynn; Jones, Craig H. – Research in the Schools, 2008
The No Child Left Behind Act and the accountability movement in public education caused many states to develop criterion-referenced academic achievement tests. Scores from these tests are often used to make high stakes decisions. Even so, these tests typically do not receive independent psychometric scrutiny. We evaluated the 2005 Arkansas…
Descriptors: Criterion Referenced Tests, Achievement Tests, High Stakes Tests, Public Education
Peer reviewedSwaminathan, Hariharan; And Others – Journal of Educational Measurement, 1974
It is proposed that the reliability of criterion-referenced test scores be defined in terms of the consistency of the decision-making process across repeated administrations of the test. (Author/RC)
Descriptors: Criterion Referenced Tests, Decision Making, Statistical Analysis, Test Reliability
Huynh, Huynh – 1977
The kappamax reliability index of domain-referenced tests is defined as the upper bound of kappa when all possibile cutoff scores are considered. Computational procedures for kappamax are described, as well as its approximation for long tests, based on Kuder-Richardson formula 21. The sampling error of kappamax, and the effects of test length and…
Descriptors: Criterion Referenced Tests, Mathematical Models, Statistical Analysis, Test Reliability
Epstein, Kenneth I. – 1977
Predictive Sample Reuse is an alternative to parameter estimation for solving a variety of statistical problems. The technique is described and its relevance to criterion referenced measurement discussed. An item analysis procedure which does not require multiple test administrations or variability in test or item scores and which adheres to the…
Descriptors: Criterion Referenced Tests, Item Analysis, Nonparametric Statistics, Predictive Measurement
Porter, Deborah Elena – 1975
Within the last decade, a wealth of literature has appeared debating the merits, use, construction, and interpretation of criterion referenced tests. This is a bibliography of articles on a wide range of topics on criterion referenced testing subjects. Also included are articles on mastery and domain referenced testing. The articles cited were…
Descriptors: Bibliographies, Criterion Referenced Tests, Elementary Secondary Education, Higher Education
Peer reviewedRozeboom, William W. – Educational and Psychological Measurement, 1978
A strict equivalence presupposed by Kaiser and Michael to derive the coefficient of "domain validity" is defensible only as a biased approximation. But then, it is far from clear what psychometric significance this coefficient has in the first place. (Author)
Descriptors: Criterion Referenced Tests, Item Analysis, Item Banks, Test Validity
Schwartz, Judah L. – Educational Technology, 1978
Professional educators as goal formulators are presented as the responsible agents of accountability. Norm- and criterion-referenced testing are compared and suggestions for accountability and goal development are given. (RAO)
Descriptors: Accountability, Criterion Referenced Tests, Educational Objectives, Norm Referenced Tests
Peer reviewedBrennan, Robert L.; Kane, Michael T. – Journal of Educational Measurement, 1977
An index for the dependability of mastery tests is described. Assumptions necessary for the index and the mathematical development of the index are provided. (Author/JKS)
Descriptors: Criterion Referenced Tests, Mastery Tests, Mathematical Models, Test Reliability
Rudner, Lawrence M. – 2000
This digest introduces ways of responding to the call for criterion-referenced information using Bayes' Theorem, a method that was coupled with criterion-referenced testing in the early 1970s (see R. Hambleton and M. Novick, 1973). To illustrate Bayes' Theorem, an example is given in which the goal is to classify an examinee as being a master or…
Descriptors: Adaptive Testing, Bayesian Statistics, Criterion Referenced Tests, Test Construction
Popham, W. James – Phi Delta Kappan, 1974
Discusses advantages and disadvantages of norm-referenced and criterion-referenced tests. (JF)
Descriptors: Criterion Referenced Tests, Norm Referenced Tests, Testing, Testing Problems
Sherrill, James L. – Educational Technology, 1972
Descriptors: Criterion Referenced Tests, Instructional Design, Material Development, Task Analysis
Deterline, William A. – NSPI Journal, 1971
A discussion of the need to utilize a training approach based on job performance requirements and criterion tests." (Author/AK)
Descriptors: Criterion Referenced Tests, Job Training, Test Construction, Testing Problems
Peer reviewedHuynh, Huynh – Journal of Educational Statistics, 1982
Two indices for assessing the efficiency of decisions in mastery testing are proposed. The indices are generalizations of the raw agreement index and the kappa index. Empirical examples of these indices are given. (Author/JKS)
Descriptors: Criterion Referenced Tests, Cutting Scores, Mastery Tests, Test Reliability
Peer reviewedBergan, John R.; And Others – Journal of Educational Statistics, 1980
Techniques for classifying individuals into mastery or non-mastery status based on criterion-referenced test performance are investigated. Extensions of mastery classification techniques developed by Macready and Dayton are presented. The techniques are based on latent class and quasi-independence models. (Author/JKS)
Descriptors: Criterion Referenced Tests, Cutting Scores, Mastery Tests, Test Interpretation

Direct link
