Publication Date
| In 2026 | 0 |
| Since 2025 | 59 |
| Since 2022 (last 5 years) | 416 |
| Since 2017 (last 10 years) | 919 |
| Since 2007 (last 20 years) | 1970 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 93 |
| Practitioners | 23 |
| Teachers | 22 |
| Policymakers | 10 |
| Administrators | 5 |
| Students | 4 |
| Counselors | 2 |
| Parents | 2 |
| Community | 1 |
Location
| United States | 47 |
| Germany | 42 |
| Australia | 34 |
| Canada | 27 |
| Turkey | 27 |
| California | 22 |
| United Kingdom (England) | 20 |
| Netherlands | 18 |
| China | 17 |
| New York | 15 |
| United Kingdom | 15 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Does not meet standards | 1 |
Peer reviewedBlair, R. Clifford; And Others – Educational and Psychological Measurement, 1983
Sampling experiments were used to assess the Type I error rates of the t test in situations where classes were randomly assigned to groups but analyses were carried out on individual student scores. Even small amounts of between-class variation caused large inflations in the Type I error rate of the t test. (Author/BW)
Descriptors: Academic Achievement, Data Analysis, Elementary Secondary Education, Error of Measurement
Peer reviewedGross, Alan L.; Kagen, Edward – Educational and Psychological Measurement, 1983
This paper compares an uncorrected with a corrected correlation between a selection test and a test-criterion in terms of expected mean square error (EMSE). It presents evidence that although the uncorrected may be more biased than the corrected correlation, it may have a smaller EMSE value, especially in small samples. (Author/PN)
Descriptors: Competitive Selection, Correlation, Error of Measurement, Research Methodology
Peer reviewedBardo, J.W.; And Others – Perceptual and Motor Skills, 1982
Data for four-, five-, and seven-position Likert formats from 292 undergraduates showed systematic error varied among formats, i.e., central tendency errors tended to increase with increasing number of categories and to reduce variances expected. (Author)
Descriptors: Error of Measurement, Higher Education, Measurement Techniques, Rating Scales
Peer reviewedWhitely, Susan E. – Applied Psychological Measurement, 1979
A model which gives maximum likelihood estimates of measurement error within the context of a simplex model for practice effects is presented. The appropriateness of the model is tested for five traits, and error estimates are compared to the classical formula estimates. (Author/JKS)
Descriptors: Error of Measurement, Error Patterns, Higher Education, Mathematical Models
Peer reviewedPursell, Elliott D.; And Others – Personnel Psychology, 1980
Results of an eight-hour training program designed to minimize rating errors supported the hypothesis that low correlations were a result of rating errors made by supervisors. Of the five predictors, four correlated significantly with the performance ratings that were conducted after the supervisors had received the training. (Author)
Descriptors: Correlation, Employees, Error of Measurement, Evaluation Criteria
Peer reviewedMitchelmore, M. C. – British Journal of Educational Psychology, 1981
This paper presents a scientific rationale for deciding the number of points to use on a grading scale in any given assessment situation. The rationale is applied to two common methods of assessment (multiple-choice and essay tests) and an example of a composite assessment. (Author/SJL)
Descriptors: Error of Measurement, Essay Tests, Grading, Higher Education
Peer reviewedChase, Christopher H.; Sattler, Jerome M. – School Psychology Review, 1980
Sattler's standard deviation technique for interpreting strengths and weaknesses on the Stanford-Binet Intelligence Scale has been simplified by Kaufman and Waterstreet in the form of an easy-to-use table. A refinement of their table is presented, with an example to demonstrate its use. (Author/CTM)
Descriptors: Chronological Age, Error of Measurement, Intelligence Tests, Mental Age
Peer reviewedMeyers, Carlton R. – Physical Educator, 1980
Salient concerns for practitioners of fitness testing are identified. (JD)
Descriptors: Error of Measurement, Exercise Physiology, Physical Fitness, Scores
Peer reviewedHays, Ron D.; Bell, Robert M.; Gillogly, James J.; Hill, Laural; Giroux, Dennis; Davis, Claude; Lewis, Matthew W.; Damush, Teresa M.; Nicholas, Ronald – Journal of Alcohol and Drug Education, 1997
Describes study results that compared different ways of measuring frequency of alcohol use employed in previous research. Results of intake interviews with clients (N=832) indicate that frequency-of-alcohol-use reports can be affected by the type of response options used and by the location of items in a self-administered microcomputer interview.…
Descriptors: Alcohol Abuse, Behavior Rating Scales, Drinking, Error of Measurement
Peer reviewedPaolo, Anthony M.; Axelrod, Bradley N.; Troster, Alexander I. – Assessment, 1996
Eighty-seven normal older people were administered the Wisconsin Sorting Test on two occasions averaging over a year apart. There were average retest gains of 5 to 7 standard score points. The standard error of prediction, standard error of difference, and abnormal test-retest discrepancy scores were calculated for clinical use. (SLD)
Descriptors: Clinical Diagnosis, Diagnostic Tests, Error of Measurement, Older Adults
Peer reviewedCullen, Rowena; Gray, Alistair – Journal of Library Administration, 1995
Due to inaccuracies in method, public libraries consistently undermeasure their reference transactions. This article chronicles the process of developing ways to systematize the count. Results of a literature survey, merits of various sample sizes, sample selection, and calculations are all discussed. Appendices include a list of categories of…
Descriptors: Error of Measurement, Foreign Countries, Library Statistics, Measurement Techniques
Peer reviewedKane, Michael – Applied Measurement in Education, 1996
This overview of the role of error and tolerance for error in measurement asserts that the generic precision associated with a measurement procedure is defined as the root mean square error, or standard error, in some relevant population. This view of precision is explored in several applications of measurement. (SLD)
Descriptors: Error of Measurement, Error Patterns, Generalizability Theory, Measurement Techniques
Peer reviewedZwick, Rebecca – Journal of Educational Statistics, 1990
Use of the Mantel-Haenszel procedure as a test for differential item functioning under the Rasch model of item-response theory is examined. Results of the procedure cannot be generalized to the class of items for which item-response functions are monotonic and local independence holds. (TJH)
Descriptors: Demography, Equations (Mathematics), Error of Measurement, Item Bias
Peer reviewedWilcox, Rand R. – Journal of Educational Statistics, 1990
Recently, C. E. McCulloch (1987) suggested a modification of the Morgan-Pitman test for comparing the variances of two dependent groups. This paper demonstrates that there are situations where the procedure is not robust. A subsample approach, similar to the Box-Scheffe test, and the Sandvik-Olsson procedure are also assessed. (TJH)
Descriptors: Comparative Analysis, Equations (Mathematics), Error of Measurement, Mathematical Models
Peer reviewedNorcini, John J.; And Others – Evaluation and the Health Professions, 1990
Aggregate scoring was applied to a recertifying examination for medical professionals to generate an answer key and allow comparison of peer examinees. Results for 1,927 candidates for recertification indicate considerable agreement between the traditional answer key and the aggregate answer key. (TJH)
Descriptors: Answer Keys, Criterion Referenced Tests, Error of Measurement, Generalizability Theory


