Publication Date
| In 2026 | 0 |
| Since 2025 | 1 |
| Since 2022 (last 5 years) | 25 |
| Since 2017 (last 10 years) | 72 |
| Since 2007 (last 20 years) | 193 |
Descriptor
| Comparative Analysis | 318 |
| Monte Carlo Methods | 318 |
| Statistical Analysis | 82 |
| Sample Size | 71 |
| Item Response Theory | 70 |
| Simulation | 65 |
| Correlation | 61 |
| Computation | 60 |
| Error of Measurement | 60 |
| Models | 56 |
| Regression (Statistics) | 43 |
| More ▼ | |
Source
Author
Publication Type
| Journal Articles | 250 |
| Reports - Research | 193 |
| Reports - Evaluative | 95 |
| Speeches/Meeting Papers | 41 |
| Dissertations/Theses -… | 14 |
| Reports - Descriptive | 8 |
| Opinion Papers | 2 |
| Guides - Non-Classroom | 1 |
| Information Analyses | 1 |
Education Level
| Higher Education | 12 |
| Elementary Education | 10 |
| Postsecondary Education | 8 |
| Secondary Education | 8 |
| Middle Schools | 7 |
| Grade 4 | 5 |
| Junior High Schools | 5 |
| Intermediate Grades | 4 |
| Grade 8 | 3 |
| High Schools | 3 |
| Grade 6 | 2 |
| More ▼ | |
Audience
| Researchers | 6 |
| Practitioners | 1 |
Location
| Austria | 3 |
| Germany | 3 |
| Australia | 2 |
| Belgium | 2 |
| Canada | 2 |
| China (Shanghai) | 2 |
| Estonia | 2 |
| France | 2 |
| Hong Kong | 2 |
| North Carolina | 2 |
| Slovakia | 2 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Lee, Soo; Suh, Youngsuk – Journal of Educational Measurement, 2018
Lord's Wald test for differential item functioning (DIF) has not been studied extensively in the context of the multidimensional item response theory (MIRT) framework. In this article, Lord's Wald test was implemented using two estimation approaches, marginal maximum likelihood estimation and Bayesian Markov chain Monte Carlo estimation, to detect…
Descriptors: Item Response Theory, Sample Size, Models, Error of Measurement
Pfaffel, Andreas; Schober, Barbara; Spiel, Christiane – Practical Assessment, Research & Evaluation, 2016
A common methodological problem in the evaluation of the predictive validity of selection methods, e.g. in educational and employment selection, is that the correlation between predictor and criterion is biased. Thorndike's (1949) formulas are commonly used to correct for this biased correlation. An alternative approach is to view the selection…
Descriptors: Comparative Analysis, Correlation, Statistical Bias, Maximum Likelihood Statistics
Olvera Astivia, Oscar L.; Zumbo, Bruno D. – Educational and Psychological Measurement, 2015
To further understand the properties of data-generation algorithms for multivariate, nonnormal data, two Monte Carlo simulation studies comparing the Vale and Maurelli method and the Headrick fifth-order polynomial method were implemented. Combinations of skewness and kurtosis found in four published articles were run and attention was…
Descriptors: Data, Simulation, Monte Carlo Methods, Comparative Analysis
Liang, Tie; Wells, Craig S. – Applied Measurement in Education, 2015
Investigating the fit of a parametric model plays a vital role in validating an item response theory (IRT) model. An area that has received little attention is the assessment of multiple IRT models used in a mixed-format test. The present study extends the nonparametric approach, proposed by Douglas and Cohen (2001), to assess model fit of three…
Descriptors: Nonparametric Statistics, Goodness of Fit, Item Response Theory, Test Format
Jin, Ying; Eason, Hershel – Journal of Educational Issues, 2016
The effects of mean ability difference (MAD) and short tests on the performance of various DIF methods have been studied extensively in previous simulation studies. Their effects, however, have not been studied under multilevel data structure. MAD was frequently observed in large-scale cross-country comparison studies where the primary sampling…
Descriptors: Test Bias, Simulation, Hierarchical Linear Modeling, Comparative Analysis
Kang, Yoonjeong; Harring, Jeffrey R.; Li, Ming – Journal of Experimental Education, 2015
The authors performed a Monte Carlo simulation to empirically investigate the robustness and power of 4 methods in testing mean differences for 2 independent groups under conditions in which 2 populations may not demonstrate the same pattern of nonnormality. The approaches considered were the t test, Wilcoxon rank-sum test, Welch-James test with…
Descriptors: Comparative Analysis, Monte Carlo Methods, Statistical Analysis, Robustness (Statistics)
Frick, Hannah; Strobl, Carolin; Zeileis, Achim – Educational and Psychological Measurement, 2015
Rasch mixture models can be a useful tool when checking the assumption of measurement invariance for a single Rasch model. They provide advantages compared to manifest differential item functioning (DIF) tests when the DIF groups are only weakly correlated with the manifest covariates available. Unlike in single Rasch models, estimation of Rasch…
Descriptors: Item Response Theory, Test Bias, Comparative Analysis, Scores
Maeda, Hotaka; Zhang, Bo – International Journal of Testing, 2017
The omega (?) statistic is reputed to be one of the best indices for detecting answer copying on multiple choice tests, but its performance relies on the accurate estimation of copier ability, which is challenging because responses from the copiers may have been contaminated. We propose an algorithm that aims to identify and delete the suspected…
Descriptors: Cheating, Test Items, Mathematics, Statistics
Asún, Rodrigo A.; Rdz-Navarro, Karina; Alvarado, Jesús M. – Sociological Methods & Research, 2016
This study compares the performance of two approaches in analysing four-point Likert rating scales with a factorial model: the classical factor analysis (FA) and the item factor analysis (IFA). For FA, maximum likelihood and weighted least squares estimations using Pearson correlation matrices among items are compared. For IFA, diagonally weighted…
Descriptors: Likert Scales, Item Analysis, Factor Analysis, Comparative Analysis
Casabianca, Jodi M.; Lewis, Charles – Journal of Educational and Behavioral Statistics, 2015
Loglinear smoothing (LLS) estimates the latent trait distribution while making fewer assumptions about its form and maintaining parsimony, thus leading to more precise item response theory (IRT) item parameter estimates than standard marginal maximum likelihood (MML). This article provides the expectation-maximization algorithm for MML estimation…
Descriptors: Item Response Theory, Maximum Likelihood Statistics, Computation, Comparative Analysis
Jan, Show-Li; Shieh, Gwowen – Journal of Educational and Behavioral Statistics, 2014
The analysis of variance (ANOVA) is one of the most frequently used statistical analyses in practical applications. Accordingly, the single and multiple comparison procedures are frequently applied to assess the differences among mean effects. However, the underlying assumption of homogeneous variances may not always be tenable. This study…
Descriptors: Sample Size, Statistical Analysis, Computation, Probability
Sen, Sedat – International Journal of Testing, 2018
Recent research has shown that over-extraction of latent classes can be observed in the Bayesian estimation of the mixed Rasch model when the distribution of ability is non-normal. This study examined the effect of non-normal ability distributions on the number of latent classes in the mixed Rasch model when estimated with maximum likelihood…
Descriptors: Item Response Theory, Comparative Analysis, Computation, Maximum Likelihood Statistics
Pokropek, Artur – Journal of Educational and Behavioral Statistics, 2016
A response model that is able to detect guessing behaviors and produce unbiased estimates in low-stake conditions using timing information is proposed. The model is a special case of the grade of membership model in which responses are modeled as partial members of a class that is affected by motivation and a class that responds only according to…
Descriptors: Reaction Time, Models, Guessing (Tests), Computation
Devlieger, Ines; Mayer, Axel; Rosseel, Yves – Educational and Psychological Measurement, 2016
In this article, an overview is given of four methods to perform factor score regression (FSR), namely regression FSR, Bartlett FSR, the bias avoiding method of Skrondal and Laake, and the bias correcting method of Croon. The bias correcting method is extended to include a reliable standard error. The four methods are compared with each other and…
Descriptors: Regression (Statistics), Comparative Analysis, Structural Equation Models, Monte Carlo Methods
McNeish, Daniel M. – Journal of Educational and Behavioral Statistics, 2016
Mixed-effects models (MEMs) and latent growth models (LGMs) are often considered interchangeable save the discipline-specific nomenclature. Software implementations of these models, however, are not interchangeable, particularly with small sample sizes. Restricted maximum likelihood estimation that mitigates small sample bias in MEMs has not been…
Descriptors: Models, Statistical Analysis, Hierarchical Linear Modeling, Sample Size

Peer reviewed
Direct link
