NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 20 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Trierweiler, Tammy J.; Lewis, Charles; Smith, Robert L. – Journal of Educational Measurement, 2016
In this study, we describe what factors influence the observed score correlation between an (external) anchor test and a total test. We show that the anchor to full-test observed score correlation is based on two components: the true score correlation between the anchor and total test, and the reliability of the anchor test. Findings using an…
Descriptors: Scores, Correlation, Tests, Test Reliability
Peer reviewed Peer reviewed
Lord, Frederic M. – Journal of Educational Measurement, 1974
Descriptors: Statistical Analysis, Test Reliability, Transformations (Mathematics)
Peer reviewed Peer reviewed
Subkoviak, Michael J. – Journal of Educational Measurement, 1988
Current methods for obtaining reliability indices for mastery tests can be laborious. This paper offers practitioners tables from which agreement and kappa coefficients can be read directly and provides criterion for acceptable values of agreement and kappa coefficients. (TJH)
Descriptors: Mastery Tests, Statistical Analysis, Test Reliability, Testing
Peer reviewed Peer reviewed
Swaminathan, Hariharan; And Others – Journal of Educational Measurement, 1974
It is proposed that the reliability of criterion-referenced test scores be defined in terms of the consistency of the decision-making process across repeated administrations of the test. (Author/RC)
Descriptors: Criterion Referenced Tests, Decision Making, Statistical Analysis, Test Reliability
Peer reviewed Peer reviewed
Gardner, P. L. – Journal of Educational Measurement, 1970
Descriptors: Error of Measurement, Mathematical Models, Statistical Analysis, Test Reliability
Peer reviewed Peer reviewed
Stafford, Richard E. – Journal of Educational Measurement, 1971
Descriptors: Correlation, Statistical Analysis, Test Interpretation, Test Reliability
Peer reviewed Peer reviewed
Lord, Frederic M. – Journal of Educational Measurement, 1974
When comparing two tests that measure the same trait, separate comparisons should be made at different levels of the trait. A simple, practical, approximate formula is given for doing this. The adequacy of the approximation is illustrated using data comparing seven nationally known sixth-grade reading tests. (Author/RC)
Descriptors: Ability Identification, Comparative Analysis, Reading Tests, Statistical Analysis
Peer reviewed Peer reviewed
Callender, John C.; Osburn, H. G. – Journal of Educational Measurement, 1979
Some procedures for estimating internal consistency reliability may be superior mathematically to the more commonly used methods such as Coefficient Alpha. One problem is computational difficulty; the other is the possibility of overestimation due to capitalization on chance. (Author/CTM)
Descriptors: Higher Education, Mathematical Formulas, Research Problems, Sampling
Peer reviewed Peer reviewed
Peng, Chao-Ying, J.; Subkoviak, Michael J. – Journal of Educational Measurement, 1980
Huynh (1976) suggested a method of approximating the reliability coefficient of a mastery test. The present study examines the accuracy of Huynh's approximation and also describes a computationally simpler approximation which appears to be generally more accurate than the former. (Author/RL)
Descriptors: Error of Measurement, Mastery Tests, Mathematical Models, Statistical Analysis
Peer reviewed Peer reviewed
Terwilliger, James S.; Lele, Kaustubh – Journal of Educational Measurement, 1979
Different indices for the internal consistency, reproducibility, or homogeneity of a test are based upon highly similar conceptual frameworks. Illustrations are presented to demonstrate how the maximum and minimum values of KR20 are influenced by test difficulty and the shape of the distribution of test scores. (Author/CTM)
Descriptors: Difficulty Level, Item Analysis, Mathematical Formulas, Statistical Analysis
Peer reviewed Peer reviewed
Beuchert, A. Kent; Mendoza, Jorge L. – Journal of Educational Measurement, 1979
Ten item discrimination indices, across a variety of item analysis situations, were compared, based on the validities of tests constructed by using each of the indices to select 40 items from a 100-item pool. Item score data were generated by a computer program and included a simulation of guessing. (Author/CTM)
Descriptors: Item Analysis, Simulation, Statistical Analysis, Test Construction
Peer reviewed Peer reviewed
Rowley, Glenn – Journal of Educational Measurement, 1978
The reliabilities of various observational measures were determined, and the influence of both the number and the length of the observation periods on reliability was examined, both separately and jointly. A single simplifying assumption leads to a variant of the Spearman-Brown formula, which may have wider application. (Author/CTM)
Descriptors: Career Development, Classroom Observation Techniques, Observation, Reliability
Peer reviewed Peer reviewed
Winne, Philip H.; Belfry, M. Joan – Journal of Educational Measurement, 1982
This review of issues about correcting for attenuation concludes that the basic difficulty lies in being able to identify and equate sources of variance in estimates of validity and reliability. Recommendations are proposed for cautious use of correction for attenuation. (Author/CM)
Descriptors: Correlation, Error of Measurement, Research Methodology, Statistical Analysis
Peer reviewed Peer reviewed
Rindler, Susan Ellerin – Journal of Educational Measurement, 1979
A sample of the literature on test speededness is reviewed; methods of assessing speededness are presented and criticized; the assumptions that underlie these methods are questioned, and alternate, multiple-administration methods are suggested. The importance of the effect of time limits is discussed. (Author/CTM)
Descriptors: Literature Reviews, Measurement Techniques, Reaction Time, Statistical Analysis
Peer reviewed Peer reviewed
Berk, Ronald A. – Journal of Educational Measurement, 1980
A dozen different approaches that yield 13 reliability indices for criterion-referenced tests were identified and grouped into three categories: threshold loss function, squared-error loss function, and domain score estimation. Indices were evaluated within each category. (Author/RL)
Descriptors: Classification, Criterion Referenced Tests, Cutting Scores, Evaluation Methods
Previous Page | Next Page ยป
Pages: 1  |  2