Descriptor
Source
| Journal of Educational… | 15 |
Author
Publication Type
| Journal Articles | 14 |
| Reports - Research | 7 |
| Reports - Evaluative | 6 |
| Information Analyses | 1 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
| National Assessment of… | 1 |
What Works Clearinghouse Rating
Peer reviewedReise, Steve P.; Yu, Jiayuan – Journal of Educational Measurement, 1990
Parameter recovery in the graded-response model was investigated using the MULTILOG computer program under default conditions. Results from 36 simulated data sets suggest that at least 500 examinees are needed to achieve adequate calibration under the graded model. Sample size had little influence on the true ability parameter's recovery. (SLD)
Descriptors: Computer Assisted Testing, Computer Simulation, Computer Software, Estimation (Mathematics)
Peer reviewedReckase, Mark D.; And Others – Journal of Educational Measurement, 1988
It is demonstrated, theoretically and empirically, that item sets can be selected that meet the unidimensionality assumption of most item response theory models, even though they require more than one ability for a correct response. A method for identifying such item sets for test development purposes is presented. (SLD)
Descriptors: Computer Simulation, Item Analysis, Latent Trait Theory, Mathematical Models
Peer reviewedBarcikowski, Robert S. – Journal of Educational Measurement, 1972
These results indicate that in deciding on the data-gathering design to be used in seeking norm information, attention should be given to item characteristics and test length with particular attention paid to the range of biserial correlations between item response and ability. (Author)
Descriptors: Item Sampling, Mathematical Models, Measurement Techniques, Monte Carlo Methods
Peer reviewedSpray, Judith A.; Welch, Catherine J. – Journal of Educational Measurement, 1990
The effect of large, within-examinee item difficulty variability on estimates of the proportion of consistent classification of examinees into mastery categories was studied over 2 test administrations for 100 simulated examinees. The proportion of consistent classifications was adequately estimated using the technique proposed by M. Subkoviak…
Descriptors: Classification, Difficulty Level, Estimation (Mathematics), Item Response Theory
Peer reviewedNandakumar, Ratna – Journal of Educational Measurement, 1991
A statistical method, W. F. Stout's statistical test of essential unidimensionality (1990), for exploring the lack of unidimensionality in test data was studied using Monte Carlo simulations. The statistical procedure is a hypothesis test of whether the essential dimensionality is one or exceeds one, regardless of the traditional dimensionality.…
Descriptors: Ability, Achievement Tests, Computer Simulation, Equations (Mathematics)
Peer reviewedGressard, Risa P.; Loyd, Brenda H. – Journal of Educational Measurement, 1991
A Monte Carlo study, which simulated 10,000 examinees' responses to four tests, investigated the effect of item stratification on parameter estimation in multiple matrix sampling of achievement data. Practical multiple matrix sampling is based on item stratification by item discrimination and a sampling plan with moderate number of subtests. (SLD)
Descriptors: Achievement Tests, Comparative Testing, Computer Simulation, Estimation (Mathematics)
Peer reviewedvan den Bergh, Huub; Eiting, Mindert H. – Journal of Educational Measurement, 1989
A method of assessing rater reliability via a design of overlapping rater teams is presented. Covariances or correlations of ratings can be analyzed with LISREL models. Models in which the rater reliabilities are congeneric, tau-equivalent, or parallel can be tested. Two examples based on essay ratings are presented. (TJH)
Descriptors: Analysis of Covariance, Computer Simulation, Correlation, Elementary Secondary Education
Peer reviewedNandakumar, Ratna – Journal of Educational Measurement, 1993
The phenomenon of simultaneous differential item functioning (DIF) amplification and cancellation and the role of the SIBTEST approach in detecting DIF are investigated with a variety of simulated test data. The effectiveness of SIBTEST is supported, and the implications of DIF amplification and cancellation are discussed. (SLD)
Descriptors: Computer Simulation, Elementary Secondary Education, Equal Education, Equations (Mathematics)
Peer reviewedSwaminathan, Hariharan; Rogers, H. Jane – Journal of Educational Measurement, 1990
A logistic regression model for characterizing differential item functioning (DIF) between two groups is presented. A distinction is drawn between uniform and nonuniform DIF in terms of model parameters. A statistic for testing the hypotheses of no DIF is developed, and simulation studies compare it with the Mantel-Haenszel procedure. (Author/TJH)
Descriptors: Comparative Analysis, Computer Simulation, Equations (Mathematics), Estimation (Mathematics)
Peer reviewedZwick, Rebecca – Journal of Educational Measurement, 1987
National Assessment of Educational Progress reading data were scaled using a unidimensional item response theory model. Bock's full-information factor analysis and Rosenbaum's test of unidimensionality were applied. Conclusions about unidimensionality for balanced incomplete block spiralled data were the same as for complete data. (Author/GDC)
Descriptors: Factor Analysis, Item Analysis, Latent Trait Theory, Mathematical Models
Peer reviewedPlake, Barbara S.; Kane, Michael T. – Journal of Educational Measurement, 1991
Several methods for determining a passing score on an examination from individual raters' estimates of minimal pass levels were compared through simulation. The methods used differed in the weighting estimates for each item received in the aggregation process. Reasons why the simplest procedure is most preferred are discussed. (SLD)
Descriptors: Comparative Analysis, Computer Simulation, Cutting Scores, Estimation (Mathematics)
Peer reviewedGustafsson, Jan-Eric – Journal of Educational Measurement, 1979
Computer generated data are used to show that Slinde and Linn's criticism of the usefulness of the Rasch model for equating (EJ 189 585) may have been the result of an artifact produced by the manner in which the samples were chosen in their study. (CTM)
Descriptors: Achievement Tests, Bias, College Entrance Examinations, Equated Scores
Peer reviewedHirsch, Thomas M. – Journal of Educational Measurement, 1989
Equatings were performed on both simulated and real data sets using common-examinee design and two abilities for each examinee. Results indicate that effective equating, as measured by comparability of true scores, is possible with the techniques used in this study. However, the stability of the ability estimates proved unsatisfactory. (TJH)
Descriptors: Academic Ability, College Students, Comparative Analysis, Computer Assisted Testing
Peer reviewedZwick, Rebecca; And Others – Journal of Educational Measurement, 1993
Two extensions of the Mantel Haenszel procedure that may be useful in assessing differential item functioning (DIF) are explored. Simulation results showed that, for both inferential procedures, the studied item should be included in the matching variable, as in the dichotomous case. (SLD)
Descriptors: Computer Simulation, Educational Assessment, Elementary Secondary Education, Equations (Mathematics)
Peer reviewedAckerman, Terry A. – Journal of Educational Measurement, 1992
The difference between item bias and item impact and the way they relate to item validity are discussed from a multidimensional item response theory perspective. The Mantel-Haenszel procedure and the Simultaneous Item Bias strategy are used in a Monte Carlo study to illustrate detection of item bias. (SLD)
Descriptors: Causal Models, Computer Simulation, Construct Validity, Equations (Mathematics)


