Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 4 |
| Since 2007 (last 20 years) | 12 |
Descriptor
| Accuracy | 12 |
| Comparative Analysis | 12 |
| Maximum Likelihood Statistics | 12 |
| Monte Carlo Methods | 8 |
| Models | 7 |
| Computation | 6 |
| Item Response Theory | 6 |
| Test Items | 5 |
| Correlation | 4 |
| Bayesian Statistics | 3 |
| Difficulty Level | 3 |
| More ▼ | |
Source
Author
| Monroe, Scott | 2 |
| Cai, Li | 1 |
| Camp-McCoy, Debra | 1 |
| Dogan, Nuri | 1 |
| Finch, Holmes | 1 |
| French, Brian F. | 1 |
| He, Wei | 1 |
| Hoaglin, David C. | 1 |
| Jiao, Hong | 1 |
| Kilic, Abdullah Faruk | 1 |
| Koziol, Natalie A. | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 12 |
| Reports - Research | 11 |
| Reports - Evaluative | 1 |
Education Level
| Higher Education | 1 |
| Postsecondary Education | 1 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Kilic, Abdullah Faruk; Dogan, Nuri – International Journal of Assessment Tools in Education, 2021
Weighted least squares (WLS), weighted least squares mean-and-variance-adjusted (WLSMV), unweighted least squares mean-and-variance-adjusted (ULSMV), maximum likelihood (ML), robust maximum likelihood (MLR) and Bayesian estimation methods were compared in mixed item response type data via Monte Carlo simulation. The percentage of polytomous items,…
Descriptors: Factor Analysis, Computation, Least Squares Statistics, Maximum Likelihood Statistics
Monroe, Scott – Journal of Educational and Behavioral Statistics, 2019
In item response theory (IRT) modeling, the Fisher information matrix is used for numerous inferential procedures such as estimating parameter standard errors, constructing test statistics, and facilitating test scoring. In principal, these procedures may be carried out using either the expected information or the observed information. However, in…
Descriptors: Item Response Theory, Error of Measurement, Scoring, Inferences
Finch, Holmes; French, Brian F. – Applied Measurement in Education, 2019
The usefulness of item response theory (IRT) models depends, in large part, on the accuracy of item and person parameter estimates. For the standard 3 parameter logistic model, for example, these parameters include the item parameters of difficulty, discrimination, and pseudo-chance, as well as the person ability parameter. Several factors impact…
Descriptors: Item Response Theory, Accuracy, Test Items, Difficulty Level
Tal, Yael; Kukliansky, Ida – Journal of Statistics Education, 2020
The aim of this study is to explore the judgments and reasoning in probabilistic tasks that require comparing two probabilities either with or without introducing an additional degree of uncertainty. The reasoning associated with the task having an additional condition of uncertainty has not been discussed in previous studies. The 66 undergraduate…
Descriptors: Undergraduate Students, Comparative Analysis, Statistics, Probability
Pfaffel, Andreas; Schober, Barbara; Spiel, Christiane – Practical Assessment, Research & Evaluation, 2016
A common methodological problem in the evaluation of the predictive validity of selection methods, e.g. in educational and employment selection, is that the correlation between predictor and criterion is biased. Thorndike's (1949) formulas are commonly used to correct for this biased correlation. An alternative approach is to view the selection…
Descriptors: Comparative Analysis, Correlation, Statistical Bias, Maximum Likelihood Statistics
Monroe, Scott; Cai, Li – Educational and Psychological Measurement, 2014
In Ramsay curve item response theory (RC-IRT) modeling, the shape of the latent trait distribution is estimated simultaneously with the item parameters. In its original implementation, RC-IRT is estimated via Bock and Aitkin's EM algorithm, which yields maximum marginal likelihood estimates. This method, however, does not produce the…
Descriptors: Item Response Theory, Models, Computation, Mathematics
Koziol, Natalie A. – Applied Measurement in Education, 2016
Testlets, or groups of related items, are commonly included in educational assessments due to their many logistical and conceptual advantages. Despite their advantages, testlets introduce complications into the theory and practice of educational measurement. Responses to items within a testlet tend to be correlated even after controlling for…
Descriptors: Classification, Accuracy, Comparative Analysis, Models
Trikalinos, Thomas A.; Hoaglin, David C.; Small, Kevin M.; Terrin, Norma; Schmid, Christopher H. – Research Synthesis Methods, 2014
Existing methods for meta-analysis of diagnostic test accuracy focus primarily on a single index test. We propose models for the joint meta-analysis of studies comparing multiple index tests on the same participants in paired designs. These models respect the grouping of data by studies, account for the within-study correlation between the tests'…
Descriptors: Meta Analysis, Diagnostic Tests, Accuracy, Comparative Analysis
Sun, Shuyan; Pan, Wei – Journal of Experimental Education, 2013
Regression discontinuity design is an alternative to randomized experiments to make causal inference when random assignment is not possible. This article first presents the formal identification and estimation of regression discontinuity treatment effects in the framework of Rubin's causal model, followed by a thorough literature review of…
Descriptors: Regression (Statistics), Computation, Accuracy, Causal Models
Camp-McCoy, Debra – Journal of the American Academy of Special Education Professionals, 2012
The Ability-Achievement Model is reviewed for efficacy in comparison to the Response to Intervention Model when diagnosing students with possible learning disabilities. The research will address the success of Ability-Achievement Model versus the Response to Intervention Model when successfully diagnosing students with learning disabilities was…
Descriptors: Response to Intervention, Models, Learning Disabilities, Diagnostic Tests
Wang, Zhen; Yao, Lihua – ETS Research Report Series, 2013
The current study used simulated data to investigate the properties of a newly proposed method (Yao's rater model) for modeling rater severity and its distribution under different conditions. Our study examined the effects of rater severity, distributions of rater severity, the difference between item response theory (IRT) models with rater effect…
Descriptors: Test Format, Test Items, Responses, Computation
Jiao, Hong; Wang, Shudong; He, Wei – Journal of Educational Measurement, 2013
This study demonstrated the equivalence between the Rasch testlet model and the three-level one-parameter testlet model and explored the Markov Chain Monte Carlo (MCMC) method for model parameter estimation in WINBUGS. The estimation accuracy from the MCMC method was compared with those from the marginalized maximum likelihood estimation (MMLE)…
Descriptors: Computation, Item Response Theory, Models, Monte Carlo Methods

Peer reviewed
Direct link
