Publication Date
| In 2026 | 0 |
| Since 2025 | 12 |
| Since 2022 (last 5 years) | 56 |
| Since 2017 (last 10 years) | 108 |
| Since 2007 (last 20 years) | 255 |
Descriptor
| Sample Size | 404 |
| Simulation | 404 |
| Item Response Theory | 113 |
| Statistical Analysis | 92 |
| Error of Measurement | 86 |
| Models | 84 |
| Test Items | 80 |
| Comparative Analysis | 77 |
| Monte Carlo Methods | 75 |
| Correlation | 67 |
| Evaluation Methods | 64 |
| More ▼ | |
Source
Author
| Fan, Xitao | 7 |
| Beretvas, S. Natasha | 5 |
| Algina, James | 4 |
| Chan, Wai | 4 |
| Cohen, Allan S. | 4 |
| De Champlain, Andre | 4 |
| Finch, W. Holmes | 4 |
| French, Brian F. | 4 |
| Kim, Seock-Ho | 4 |
| Kromrey, Jeffrey D. | 4 |
| Paek, Insu | 4 |
| More ▼ | |
Publication Type
Education Level
Audience
| Teachers | 4 |
| Researchers | 3 |
Location
| North Carolina | 2 |
| Armenia | 1 |
| Austria | 1 |
| Canada | 1 |
| Florida (Miami) | 1 |
| Hong Kong | 1 |
| Indiana | 1 |
| Iran | 1 |
| Montana | 1 |
| New York (New York) | 1 |
| Norway | 1 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 2 |
Assessments and Surveys
What Works Clearinghouse Rating
Kabasakal, Kübra Atalay; Kelecioglu, Hülya – Educational Sciences: Theory and Practice, 2015
This study examines the effect of differential item functioning (DIF) items on test equating through multilevel item response models (MIRMs) and traditional IRMs. The performances of three different equating models were investigated under 24 different simulation conditions, and the variables whose effects were examined included sample size, test…
Descriptors: Test Bias, Equated Scores, Item Response Theory, Simulation
Tipton, Elizabeth; Pustejovsky, James E. – Journal of Educational and Behavioral Statistics, 2015
Meta-analyses often include studies that report multiple effect sizes based on a common pool of subjects or that report effect sizes from several samples that were treated with very similar research protocols. The inclusion of such studies introduces dependence among the effect size estimates. When the number of studies is large, robust variance…
Descriptors: Meta Analysis, Effect Size, Computation, Robustness (Statistics)
McNeish, Daniel – Review of Educational Research, 2017
In education research, small samples are common because of financial limitations, logistical challenges, or exploratory studies. With small samples, statistical principles on which researchers rely do not hold, leading to trust issues with model estimates and possible replication issues when scaling up. Researchers are generally aware of such…
Descriptors: Models, Statistical Analysis, Sampling, Sample Size
Sahin, Alper; Weiss, David J. – Educational Sciences: Theory and Practice, 2015
This study aimed to investigate the effects of calibration sample size and item bank size on examinee ability estimation in computerized adaptive testing (CAT). For this purpose, a 500-item bank pre-calibrated using the three-parameter logistic model with 10,000 examinees was simulated. Calibration samples of varying sizes (150, 250, 350, 500,…
Descriptors: Adaptive Testing, Computer Assisted Testing, Sample Size, Item Banks
Oshima, T. C.; Wright, Keith; White, Nick – International Journal of Testing, 2015
Raju, van der Linden, and Fleer (1995) introduced a framework for differential functioning of items and tests (DFIT) for unidimensional dichotomous models. Since then, DFIT has been shown to be a quite versatile framework as it can handle polytomous as well as multidimensional models both at the item and test levels. However, DFIT is still limited…
Descriptors: Test Bias, Item Response Theory, Test Items, Simulation
Babcock, Ben; Albano, Anthony; Raymond, Mark – Educational and Psychological Measurement, 2012
The authors introduced nominal weights mean equating, a simplified version of Tucker equating, as an alternative for dealing with very small samples. The authors then conducted three simulation studies to compare nominal weights mean equating to six other equating methods under the nonequivalent groups anchor test design with sample sizes of 20,…
Descriptors: Equated Scores, Methods, Sample Size, Simulation
Lee, Hollylynne S.; Starling, Tina T.; Gonzalez, Marggie D. – Mathematics Teacher, 2014
Research shows that students often struggle with understanding empirical sampling distributions. Using hands-on and technology models and simulations of problems generated by real data help students begin to make connections between repeated sampling, sample size, distribution, variation, and center. A task to assist teachers in implementing…
Descriptors: Sampling, Sample Size, Statistical Distributions, Simulation
Beasley, T. Mark – Journal of Experimental Education, 2014
Increasing the correlation between the independent variable and the mediator ("a" coefficient) increases the effect size ("ab") for mediation analysis; however, increasing a by definition increases collinearity in mediation models. As a result, the standard error of product tests increase. The variance inflation caused by…
Descriptors: Statistical Analysis, Effect Size, Nonparametric Statistics, Statistical Inference
de Winter, J. C .F. – Practical Assessment, Research & Evaluation, 2013
Researchers occasionally have to work with an extremely small sample size, defined herein as "N" less than or equal to 5. Some methodologists have cautioned against using the "t"-test when the sample size is extremely small, whereas others have suggested that using the "t"-test is feasible in such a case. The present…
Descriptors: Sample Size, Statistical Analysis, Hypothesis Testing, Simulation
Lamsal, Sunil – ProQuest LLC, 2015
Different estimation procedures have been developed for the unidimensional three-parameter item response theory (IRT) model. These techniques include the marginal maximum likelihood estimation, the fully Bayesian estimation using Markov chain Monte Carlo simulation techniques, and the Metropolis-Hastings Robbin-Monro estimation. With each…
Descriptors: Item Response Theory, Monte Carlo Methods, Maximum Likelihood Statistics, Markov Processes
Schoeneberger, Jason A. – Journal of Experimental Education, 2016
The design of research studies utilizing binary multilevel models must necessarily incorporate knowledge of multiple factors, including estimation method, variance component size, or number of predictors, in addition to sample sizes. This Monte Carlo study examined the performance of random effect binary outcome multilevel models under varying…
Descriptors: Sample Size, Models, Computation, Predictor Variables
Rusticus, Shayna A.; Lovato, Chris Y. – Practical Assessment, Research & Evaluation, 2014
The question of equivalence between two or more groups is frequently of interest to many applied researchers. Equivalence testing is a statistical method designed to provide evidence that groups are comparable by demonstrating that the mean differences found between groups are small enough that they are considered practically unimportant. Few…
Descriptors: Sample Size, Equivalency Tests, Simulation, Error of Measurement
Zwick, Rebecca; Ye, Lei; Isham, Steven – ETS Research Report Series, 2013
Differential item functioning (DIF) analysis is a key component in the evaluation of the fairness and validity of educational tests. Although it is often assumed that refinement of the matching criterion always provides more accurate DIF results, the actual situation proves to be more complex. To explore the effectiveness of refinement, we…
Descriptors: Test Bias, Statistical Analysis, Simulation, Educational Testing
Lathrop, Quinn N.; Cheng, Ying – Applied Psychological Measurement, 2013
Within the framework of item response theory (IRT), there are two recent lines of work on the estimation of classification accuracy (CA) rate. One approach estimates CA when decisions are made based on total sum scores, the other based on latent trait estimates. The former is referred to as the Lee approach, and the latter, the Rudner approach,…
Descriptors: Item Response Theory, Accuracy, Classification, Computation
Keller, Bryan – Psychometrika, 2012
Randomization tests are often recommended when parametric assumptions may be violated because they require no distributional or random sampling assumptions in order to be valid. In addition to being exact, a randomization test may also be more powerful than its parametric counterpart. This was demonstrated in a simulation study which examined the…
Descriptors: Statistical Analysis, Nonparametric Statistics, Simulation, Sampling

Peer reviewed
Direct link
