Publication Date
| In 2026 | 0 |
| Since 2025 | 12 |
| Since 2022 (last 5 years) | 56 |
| Since 2017 (last 10 years) | 108 |
| Since 2007 (last 20 years) | 255 |
Descriptor
| Sample Size | 404 |
| Simulation | 404 |
| Item Response Theory | 113 |
| Statistical Analysis | 92 |
| Error of Measurement | 86 |
| Models | 84 |
| Test Items | 80 |
| Comparative Analysis | 77 |
| Monte Carlo Methods | 75 |
| Correlation | 67 |
| Evaluation Methods | 64 |
| More ▼ | |
Source
Author
| Fan, Xitao | 7 |
| Beretvas, S. Natasha | 5 |
| Algina, James | 4 |
| Chan, Wai | 4 |
| Cohen, Allan S. | 4 |
| De Champlain, Andre | 4 |
| Finch, W. Holmes | 4 |
| French, Brian F. | 4 |
| Kim, Seock-Ho | 4 |
| Kromrey, Jeffrey D. | 4 |
| Paek, Insu | 4 |
| More ▼ | |
Publication Type
Education Level
Audience
| Teachers | 4 |
| Researchers | 3 |
Location
| North Carolina | 2 |
| Armenia | 1 |
| Austria | 1 |
| Canada | 1 |
| Florida (Miami) | 1 |
| Hong Kong | 1 |
| Indiana | 1 |
| Iran | 1 |
| Montana | 1 |
| New York (New York) | 1 |
| Norway | 1 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 2 |
Assessments and Surveys
What Works Clearinghouse Rating
Chan, Wendy – Journal of Educational and Behavioral Statistics, 2018
Policymakers have grown increasingly interested in how experimental results may generalize to a larger population. However, recently developed propensity score-based methods are limited by small sample sizes, where the experimental study is generalized to a population that is at least 20 times larger. This is particularly problematic for methods…
Descriptors: Computation, Generalization, Probability, Sample Size
McNeish, Daniel – Journal of Experimental Education, 2018
Small samples are common in growth models due to financial and logistical difficulties of following people longitudinally. For similar reasons, longitudinal studies often contain missing data. Though full information maximum likelihood (FIML) is popular to accommodate missing data, the limited number of studies in this area have found that FIML…
Descriptors: Growth Models, Sampling, Sample Size, Hierarchical Linear Modeling
Craig, Brandon – ProQuest LLC, 2017
The purpose of this study was to determine if using a multistage approach for the empirical selection of anchor items would lead to more accurate DIF detection rates than the anchor selection methods proposed by Kopf, Zeileis, & Strobl (2015b). A simulation study was conducted in which the sample size, percentage of DIF, and balance of DIF…
Descriptors: Simulation, Sample Size, Item Response Theory, Item Analysis
Choi, In-Hee; Paek, Insu; Cho, Sun-Joo – Journal of Experimental Education, 2017
The purpose of the current study is to examine the performance of four information criteria (Akaike's information criterion [AIC], corrected AIC [AICC] Bayesian information criterion [BIC], sample-size adjusted BIC [SABIC]) for detecting the correct number of latent classes in the mixture Rasch model through simulations. The simulation study…
Descriptors: Item Response Theory, Models, Bayesian Statistics, Simulation
Rutkowski, Leslie; Svetina, Dubravka – Applied Measurement in Education, 2017
In spite of the challenges inherent in making dozens of comparisons across heterogeneous populations, a relatively recent interest in scale-score equivalence for non-achievement measures in an international context has emerged. Until recently, operational procedures for establishing measurement invariance using multiple-groups analyses were…
Descriptors: International Assessment, Goodness of Fit, Statistical Analysis, Teacher Surveys
Kang, Yoonjeong; McNeish, Daniel M.; Hancock, Gregory R. – Educational and Psychological Measurement, 2016
Although differences in goodness-of-fit indices (?GOFs) have been advocated for assessing measurement invariance, studies that advanced recommended differential cutoffs for adjudicating invariance actually utilized a very limited range of values representing the quality of indicator variables (i.e., magnitude of loadings). Because quality of…
Descriptors: Measurement, Goodness of Fit, Guidelines, Models
Asiret, Semih; Sünbül, Seçil Ömür – Educational Sciences: Theory and Practice, 2016
In this study, equating methods for random group design using small samples through factors such as sample size, difference in difficulty between forms, and guessing parameter was aimed for comparison. Moreover, which method gives better results under which conditions was also investigated. In this study, 5,000 dichotomous simulated data…
Descriptors: Equated Scores, Sample Size, Difficulty Level, Guessing (Tests)
Liu, Jin – ProQuest LLC, 2015
Statistical power is important in a meta-analysis study, although few studies have examined the performance of simulated power in meta-analysis. The purpose of this study is to inform researchers about statistical power estimation on two sample mean difference test under different situations: (1) the discrepancy between the analytical power and…
Descriptors: Statistical Analysis, Meta Analysis, Simulation, Computation
Kang, Hyeon-Ah; Lu, Ying; Chang, Hua-Hua – Applied Measurement in Education, 2017
Increasing use of item pools in large-scale educational assessments calls for an appropriate scaling procedure to achieve a common metric among field-tested items. The present study examines scaling procedures for developing a new item pool under a spiraled block linking design. The three scaling procedures are considered: (a) concurrent…
Descriptors: Item Response Theory, Accuracy, Educational Assessment, Test Items
Morgan, Grant B.; Moore, Courtney A.; Floyd, Harlee S. – Journal of Psychoeducational Assessment, 2018
Although content validity--how well each item of an instrument represents the construct being measured--is foundational in the development of an instrument, statistical validity is also important to the decisions that are made based on the instrument. The primary purpose of this study is to demonstrate how simulation studies can be used to assist…
Descriptors: Simulation, Decision Making, Test Construction, Validity
Bradshaw, Laine P.; Madison, Matthew J. – International Journal of Testing, 2016
In item response theory (IRT), the invariance property states that item parameter estimates are independent of the examinee sample, and examinee ability estimates are independent of the test items. While this property has long been established and understood by the measurement community for IRT models, the same cannot be said for diagnostic…
Descriptors: Classification, Models, Simulation, Psychometrics
Leventhal, Brian – ProQuest LLC, 2017
More robust and rigorous psychometric models, such as multidimensional Item Response Theory models, have been advocated for survey applications. However, item responses may be influenced by construct-irrelevant variance factors such as preferences for extreme response options. Through empirical and simulation methods, this study evaluates the use…
Descriptors: Psychometrics, Item Response Theory, Simulation, Models
Kogar, Hakan – International Journal of Assessment Tools in Education, 2018
The aim of this simulation study, determine the relationship between true latent scores and estimated latent scores by including various control variables and different statistical models. The study also aimed to compare the statistical models and determine the effects of different distribution types, response formats and sample sizes on latent…
Descriptors: Simulation, Context Effect, Computation, Statistical Analysis
Cheng, Ying; Shao, Can; Lathrop, Quinn N. – Educational and Psychological Measurement, 2016
Due to its flexibility, the multiple-indicator, multiple-causes (MIMIC) model has become an increasingly popular method for the detection of differential item functioning (DIF). In this article, we propose the mediated MIMIC model method to uncover the underlying mechanism of DIF. This method extends the usual MIMIC model by including one variable…
Descriptors: Test Bias, Models, Simulation, Sample Size
Willse, John T. – Measurement and Evaluation in Counseling and Development, 2017
This article provides a brief introduction to the Rasch model. Motivation for using Rasch analyses is provided. Important Rasch model concepts and key aspects of result interpretation are introduced, with major points reinforced using a simulation demonstration. Concrete guidelines are provided regarding sample size and the evaluation of items.
Descriptors: Item Response Theory, Test Results, Test Interpretation, Simulation

Peer reviewed
Direct link
