Descriptor
| Comparative Analysis | 12 |
| Simulation | 12 |
| Statistical Studies | 12 |
| Test Items | 6 |
| Research Methodology | 5 |
| Item Response Theory | 4 |
| Selection | 4 |
| Factor Analysis | 3 |
| Goodness of Fit | 3 |
| Models | 3 |
| Multivariate Analysis | 3 |
| More ▼ | |
Source
| Educational and Psychological… | 2 |
| Applied Psychological… | 1 |
| Journal of Educational… | 1 |
| Multivariate Behavioral… | 1 |
Author
Publication Type
| Reports - Research | 7 |
| Speeches/Meeting Papers | 7 |
| Journal Articles | 5 |
| Reports - Evaluative | 5 |
Education Level
Audience
| Researchers | 2 |
Location
Laws, Policies, & Programs
Assessments and Surveys
| ACT Assessment | 1 |
What Works Clearinghouse Rating
Peer reviewedMarcoulides, George A. – Educational and Psychological Measurement, 1994
Effects of different weighting schemes on selecting the optimal number of observations in multivariate-multifacet generalizability designs are studied when cost constraints are imposed. Comparison of four schemes through simulation indicates that all four produce similar optimal values and that reliability should be similar. (SLD)
Descriptors: Budgeting, Comparative Analysis, Costs, Factor Analysis
Peer reviewedSmith, Richard M. – Educational and Psychological Measurement, 1994
Rasch model total-fit statistics and between-item fit statistics were compared for their ability to detect measurement disturbances through the use of simulated data. Results indicate that the between-fit statistic appears more sensitive to systematic measurement disturbances and the total-fit statistic is more sensitive to random measurement…
Descriptors: Comparative Analysis, Goodness of Fit, Item Response Theory, Measurement Techniques
Peer reviewedCohen, Allan S.; And Others – Applied Psychological Measurement, 1993
Three measures of differential item functioning for the dichotomous response model are extended to include Samejima's graded response model. Two are based on area differences between item true score functions, and one is a chi-square statistic for comparing differences in item parameters. (SLD)
Descriptors: Chi Square, Comparative Analysis, Identification, Item Bias
Pommerich, Mary; And Others – 1994
The functioning of two population-based Mantel-Haenszel (MH) common-odds ratios was compared. One ratio is conditioned on the observed test score, while the other is conditioned on a latent trait or true ability score. When the comparison group distributions are incongruent or nonoverlapping to some degree, the observed score represents different…
Descriptors: Ability, Comparative Analysis, Item Bias, Performance
Peer reviewedNandakumar, Ratna – Journal of Educational Measurement, 1994
Using simulated and real data, this study compares the performance of three methodologies for assessing unidimensionality: (1) DIMTEST; (2) the approach of Holland and Rosenbaum; and (3) nonlinear factor analysis. All three models correctly confirm unidimensionality, but they differ in their ability to detect the lack of unidimensionality.…
Descriptors: Ability, Comparative Analysis, Evaluation Methods, Factor Analysis
Beasley, T. Mark; Sheehan, Janet K. – 1994
C. L. Olson (1976, 1979) suggests the Pillai-Bartlett trace (V) as an omnibus multivariate analysis of variance (MANOVA) test statistic for its superior robustness to heterogeneous variances. J. Stevens (1979, 1980) contends that the robustness of V, Wilk's lambda (W) and the Hotelling-Lawley trace (T) are similar, and that their power functions…
Descriptors: Analysis of Covariance, Comparative Analysis, Matrices, Monte Carlo Methods
Schumacker, Randall E.; And Others – 1994
Rasch between and total weighted and unweighted fit statistics were compared using varying test lengths and sample sizes. Two test lengths (20 and 50 items) and three sample sizes (150, 500, and 1,000 were crossed. Each of the six combinations were replicated 100 times. In addition, power comparisons were made. Results indicated that there were no…
Descriptors: Comparative Analysis, Goodness of Fit, Item Response Theory, Power (Statistics)
PDF pending restorationSchumacker, Randall E. – 1994
A population data set was randomly generated from which a random sample was drawn. This sample was randomly divided into two data sets, one of which was used to generate parameter estimates, which were then used in the second data set for cross-validation purposes. The best variable subset models were compared between the two data sets on the…
Descriptors: Comparative Analysis, Criteria, Estimation (Mathematics), Factor Analysis
Wang, Tianyou; Kolen, Michael J. – 1994
In this paper a quadratic curve equating method for different test forms under a random-group data-collection design is proposed. Procedures for implementing this method and related issues are described and discussed. The quadratic-curve method was evaluated with real test data (from two 30-item subtests for a professional licensure examination…
Descriptors: Comparative Analysis, Data Collection, Equated Scores, Goodness of Fit
Cope, Ronald T. – 1986
Comparisons were made of three Angoff Design V linear equating methods (two forms equated to a common test, two forms predicted by a common test, or two forms used to predict a common test) and Tucker's and R. Levine's linear methods, under common item linear equating with non-equivalent populations. Forms of a professional certification test…
Descriptors: Certification, Comparative Analysis, Equated Scores, Higher Education
Peer reviewedTang, K. Linda; Algina, James – Multivariate Behavioral Research, 1993
Type I error rates of four multivariate tests (Pilai-Bartlett trace, Johansen's test, James' first-order test, and James' second-order test) were compared for heterogeneous covariance matrices in 360 simulated experiments. The superior performance of Johansen's test and James' second-order test is discussed. (SLD)
Descriptors: Analysis of Covariance, Analysis of Variance, Comparative Analysis, Equations (Mathematics)
Sarvela, Paul D. – 1986
Four discrimination indices were compared, using score distributions which were normal, bimodal, and negatively skewed. The score distributions were systematically varied to represent the common circumstances of a military training situation using criterion-referenced mastery tests. Three 20-item tests were administered to 110 simulated subjects.…
Descriptors: Comparative Analysis, Criterion Referenced Tests, Item Analysis, Mastery Tests


