Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 5 |
Since 2006 (last 20 years) | 9 |
Descriptor
Maximum Likelihood Statistics | 17 |
Monte Carlo Methods | 17 |
Test Items | 17 |
Item Response Theory | 12 |
Bayesian Statistics | 6 |
Difficulty Level | 6 |
Markov Processes | 6 |
Simulation | 6 |
Comparative Analysis | 5 |
Computation | 5 |
Models | 5 |
More ▼ |
Source
Educational and Psychological… | 3 |
Applied Measurement in… | 2 |
Applied Psychological… | 2 |
ETS Research Report Series | 2 |
Journal of Educational… | 1 |
Journal of Educational and… | 1 |
ProQuest LLC | 1 |
Author
Cohen, Allan S. | 2 |
Finch, Holmes | 2 |
Kim, Seock-Ho | 2 |
Weiss, David J. | 2 |
Andersson, Björn | 1 |
Andreas Kurz | 1 |
Can Gürer | 1 |
Clemens Draxler | 1 |
Edwards, Julianne M. | 1 |
French, Brian F. | 1 |
He, Wei | 1 |
More ▼ |
Publication Type
Reports - Research | 12 |
Journal Articles | 11 |
Reports - Evaluative | 4 |
Speeches/Meeting Papers | 4 |
Dissertations/Theses -… | 1 |
Numerical/Quantitative Data | 1 |
Education Level
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
Comprehensive Tests of Basic… | 1 |
Test of English as a Foreign… | 1 |
What Works Clearinghouse Rating
Clemens Draxler; Andreas Kurz; Can Gürer; Jan Philipp Nolte – Journal of Educational and Behavioral Statistics, 2024
A modified and improved inductive inferential approach to evaluate item discriminations in a conditional maximum likelihood and Rasch modeling framework is suggested. The new approach involves the derivation of four hypothesis tests. It implies a linear restriction of the assumed set of probability distributions in the classical approach that…
Descriptors: Inferences, Test Items, Item Analysis, Maximum Likelihood Statistics
Finch, Holmes; French, Brian F. – Applied Measurement in Education, 2019
The usefulness of item response theory (IRT) models depends, in large part, on the accuracy of item and person parameter estimates. For the standard 3 parameter logistic model, for example, these parameters include the item parameters of difficulty, discrimination, and pseudo-chance, as well as the person ability parameter. Several factors impact…
Descriptors: Item Response Theory, Accuracy, Test Items, Difficulty Level
Andersson, Björn; Xin, Tao – Educational and Psychological Measurement, 2018
In applications of item response theory (IRT), an estimate of the reliability of the ability estimates or sum scores is often reported. However, analytical expressions for the standard errors of the estimators of the reliability coefficients are not available in the literature and therefore the variability associated with the estimated reliability…
Descriptors: Item Response Theory, Test Reliability, Test Items, Scores
Lamsal, Sunil – ProQuest LLC, 2015
Different estimation procedures have been developed for the unidimensional three-parameter item response theory (IRT) model. These techniques include the marginal maximum likelihood estimation, the fully Bayesian estimation using Markov chain Monte Carlo simulation techniques, and the Metropolis-Hastings Robbin-Monro estimation. With each…
Descriptors: Item Response Theory, Monte Carlo Methods, Maximum Likelihood Statistics, Markov Processes
Finch, Holmes; Edwards, Julianne M. – Educational and Psychological Measurement, 2016
Standard approaches for estimating item response theory (IRT) model parameters generally work under the assumption that the latent trait being measured by a set of items follows the normal distribution. Estimation of IRT parameters in the presence of nonnormal latent traits has been shown to generate biased person and item parameter estimates. A…
Descriptors: Item Response Theory, Computation, Nonparametric Statistics, Bayesian Statistics
Koziol, Natalie A. – Applied Measurement in Education, 2016
Testlets, or groups of related items, are commonly included in educational assessments due to their many logistical and conceptual advantages. Despite their advantages, testlets introduce complications into the theory and practice of educational measurement. Responses to items within a testlet tend to be correlated even after controlling for…
Descriptors: Classification, Accuracy, Comparative Analysis, Models
Seo, Dong Gi; Weiss, David J. – Educational and Psychological Measurement, 2013
The usefulness of the l[subscript z] person-fit index was investigated with achievement test data from 20 exams given to more than 3,200 college students. Results for three methods of estimating ? showed that the distributions of l[subscript z] were not consistent with its theoretical distribution, resulting in general overfit to the item response…
Descriptors: Achievement Tests, College Students, Goodness of Fit, Item Response Theory
Wang, Zhen; Yao, Lihua – ETS Research Report Series, 2013
The current study used simulated data to investigate the properties of a newly proposed method (Yao's rater model) for modeling rater severity and its distribution under different conditions. Our study examined the effects of rater severity, distributions of rater severity, the difference between item response theory (IRT) models with rater effect…
Descriptors: Test Format, Test Items, Responses, Computation
Jiao, Hong; Wang, Shudong; He, Wei – Journal of Educational Measurement, 2013
This study demonstrated the equivalence between the Rasch testlet model and the three-level one-parameter testlet model and explored the Markov Chain Monte Carlo (MCMC) method for model parameter estimation in WINBUGS. The estimation accuracy from the MCMC method was compared with those from the marginalized maximum likelihood estimation (MMLE)…
Descriptors: Computation, Item Response Theory, Models, Monte Carlo Methods

Cohen, Allan S.; And Others – Applied Psychological Measurement, 1996
Type I error rates for the likelihood ratio test for detecting differential item functioning (DIF) were investigated using Monte Carlo simulations. Type I error rates for the two-parameter model were within theoretically expected values at each alpha level, but those for the three-parameter model were not. (SLD)
Descriptors: Identification, Item Bias, Item Response Theory, Maximum Likelihood Statistics
Kim, Seock-Ho; Cohen, Allan S. – 1997
Type I error rates of the likelihood ratio test for the detection of differential item functioning (DIF) were investigated using Monte Carlo simulations. The graded response model with five ordered categories was used to generate data sets of a 30-item test for samples of 300 and 1,000 simulated examinees. All DIF comparisons were simulated by…
Descriptors: Ability, Classification, Computer Simulation, Estimation (Mathematics)

Kim, Seock-Ho; And Others – Applied Psychological Measurement, 1994
Type I error rates of F. M. Lord's chi square test for differential item functioning were investigated using Monte Carlo simulations with marginal maximum likelihood estimation and marginal Bayesian estimation algorithms. Lord's chi square did not provide useful Type I error control for the three-parameter logistic model at these sample sizes.…
Descriptors: Algorithms, Bayesian Statistics, Chi Square, Error of Measurement
Maurelli, Vincent A.; Weiss, David J. – 1981
A monte carlo simulation was conducted to assess the effects in an adaptive testing strategy for test batteries of varying subtest order, subtest termination criterion, and variable versus fixed entry on the psychometric properties of an existent achievement test battery. Comparisons were made among conventionally administered tests and adaptive…
Descriptors: Achievement Tests, Adaptive Testing, Computer Assisted Testing, Latent Trait Theory
de la Torre, Jimmy; Patz, Richard J. – 2001
This paper seeks to extend the application of Markov chain Monte Carlo (MCMC) methods in item response theory (IRT) to include the estimation of equating relationships along with the estimation of test item parameters. A method is proposed that incorporates estimation of the equating relationship in the item calibration phase. Item parameters from…
Descriptors: Achievement Tests, Bayesian Statistics, Equated Scores, Estimation (Mathematics)
Xiao, Beiling – 1990
Dichotomous search strategies (DSSs) for computerized adaptive testing are similar to golden section search strategies (GSSSs). Each middle point of successive search regions is a testing point. After each item is administered, the subject's obtained score is compared with the expected score at successive testing points. If the subject's obtained…
Descriptors: Ability Identification, Adaptive Testing, Computer Assisted Testing, Equations (Mathematics)
Previous Page | Next Page »
Pages: 1 | 2