NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 15 results Save | Export
Peer reviewed Peer reviewed
Waller, Michael I. – Journal of Educational Measurement, 1981
A method based on the likelihood ratio procedure is presented for use in selecting a measurement model from among the Rasch, two-parameter, and three-parameter logistic latent trait models. (Author/BW)
Descriptors: Comparative Analysis, Goodness of Fit, Latent Trait Theory, Mathematical Models
Peer reviewed Peer reviewed
Marascuilo, Leonard A.; Slaughter, Robert E. – Journal of Educational Measurement, 1981
Six statistical methods for identifying possible sources of bias in standardized test items are presented. The relationship between chi-squared methods and item-response theory methods are also discussed. (Author/BW)
Descriptors: Comparative Analysis, Latent Trait Theory, Mathematical Models, Standardized Tests
Peer reviewed Peer reviewed
Marsh, Herbert W.; Hocevar, Dennis – Journal of Educational Measurement, 1983
This paper describes a variety of confirmatory factor analysis models that provide improved tests of multitrait-multimethod matrices, and compares three different approaches (the original Campbell-Fiske guidelines, an analysis of variance model, and confirmatory factor analysis models). (PN)
Descriptors: Analysis of Variance, Comparative Analysis, Evaluation Methods, Factor Analysis
Peer reviewed Peer reviewed
Kane, Michael T. – Journal of Educational Measurement, 1987
The use of item response theory models for analyzing the results of judgmental standard setting studies (the Angoff technique) for establishing minimum pass levels is discussed. A comparison of three methods indicates the traditional approach may not be best. A procedure based on generalizability theory is suggested. (GDC)
Descriptors: Comparative Analysis, Cutting Scores, Generalizability Theory, Latent Trait Theory
Peer reviewed Peer reviewed
Hanson, Bradley A.; Brennan, Robert L. – Journal of Educational Measurement, 1990
Using several data sets, the relative performance of the beta binomial model and two more general strong true score models in estimating several indices of classification consistency is examined. It appears that the beta binomial model can provide inadequate fits to raw score distributions compared to more general models. (TJH)
Descriptors: Classification, Comparative Analysis, Equations (Mathematics), Estimation (Mathematics)
Peer reviewed Peer reviewed
Kolen, Michael J. – Journal of Educational Measurement, 1991
Estimation/smoothing methods that are flexible enough to fit a wide variety of test score distributions are reviewed: kernel method, strong true-score model-based method, and method that uses polynomial log-linear models. Applications of these methods include describing/comparing test score distributions, estimating norms, and estimating…
Descriptors: Comparative Analysis, Equated Scores, Equations (Mathematics), Estimation (Mathematics)
Peer reviewed Peer reviewed
Levin, Joel R. – Journal of Educational Measurement, 1975
A set procedure developed in this study is useful in determining sample size, based on specification of linear contrasts involving certain formula treatments. (Author/DEP)
Descriptors: Analysis of Variance, Comparative Analysis, Mathematical Models, Measurement Techniques
Peer reviewed Peer reviewed
Huynh, Huynh; Saunders, Joseph C. – Journal of Educational Measurement, 1980
Single administration (beta-binomial) estimates for the raw agreement index p and the corrected-for-chance kappa index in mastery testing are compared with those based on two test administrations in terms of estimation bias and sampling variability. Bias is about 2.5 percent for p and 10 percent for kappa. (Author/RL)
Descriptors: Comparative Analysis, Error of Measurement, Mastery Tests, Mathematical Models
Peer reviewed Peer reviewed
Swaminathan, Hariharan; Rogers, H. Jane – Journal of Educational Measurement, 1990
A logistic regression model for characterizing differential item functioning (DIF) between two groups is presented. A distinction is drawn between uniform and nonuniform DIF in terms of model parameters. A statistic for testing the hypotheses of no DIF is developed, and simulation studies compare it with the Mantel-Haenszel procedure. (Author/TJH)
Descriptors: Comparative Analysis, Computer Simulation, Equations (Mathematics), Estimation (Mathematics)
Peer reviewed Peer reviewed
Kolen, Michael J.; Whitney, Douglas R. – Journal of Educational Measurement, 1982
The adequacy of equipercentile, linear, one-parameter (Rasch), and three-parameter logistic item-response theory procedures for equating 12 forms of five tests of general educational development were compared. Results indicated the equating method adequacy depends on a variety of factors such as test characteristics, equating design, and sample…
Descriptors: Achievement Tests, Comparative Analysis, Equated Scores, Equivalency Tests
Peer reviewed Peer reviewed
Kim, Seock-Ho; Cohen, Allan S. – Journal of Educational Measurement, 1992
Effects of the following methods for linking metrics on detection of differential item functioning (DIF) were compared: (1) test characteristic curve method (TCC); (2) weighted mean and sigma method; and (3) minimum chi-square method. With large samples, results were essentially the same. With small samples, TCC was most accurate. (SLD)
Descriptors: Chi Square, Comparative Analysis, Equations (Mathematics), Estimation (Mathematics)
Peer reviewed Peer reviewed
Baker, Frank B.; Al-Karni, Ali – Journal of Educational Measurement, 1991
Two methods of computing test equating coefficients under item response theory by the following authors are compared: (1) B. H. Loyd and H. D. Hoover (1980); and (2) M. L. Stocking and F. M. Lord (1983). Conditions under which the method of Stocking and Lord is preferable are described. (SLD)
Descriptors: Ability, College Entrance Examinations, Comparative Analysis, Equated Scores
Peer reviewed Peer reviewed
Plake, Barbara S.; Kane, Michael T. – Journal of Educational Measurement, 1991
Several methods for determining a passing score on an examination from individual raters' estimates of minimal pass levels were compared through simulation. The methods used differed in the weighting estimates for each item received in the aggregation process. Reasons why the simplest procedure is most preferred are discussed. (SLD)
Descriptors: Comparative Analysis, Computer Simulation, Cutting Scores, Estimation (Mathematics)
Peer reviewed Peer reviewed
Marsh, Herbert W. – Journal of Educational Measurement, 1993
Structural equation models of the same construct collected on different occasions are evaluated in 2 studies involving the evaluation of 157 college instructors over 8 years and data for over 2,200 high school students over 4 years for the Youth in Transition Study. Results challenge overreliance on simplex models. (SLD)
Descriptors: College Faculty, Comparative Analysis, High School Students, High Schools
Peer reviewed Peer reviewed
Hirsch, Thomas M. – Journal of Educational Measurement, 1989
Equatings were performed on both simulated and real data sets using common-examinee design and two abilities for each examinee. Results indicate that effective equating, as measured by comparability of true scores, is possible with the techniques used in this study. However, the stability of the ability estimates proved unsatisfactory. (TJH)
Descriptors: Academic Ability, College Students, Comparative Analysis, Computer Assisted Testing