Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 1 |
| Since 2007 (last 20 years) | 2 |
Descriptor
Source
| Educational and Psychological… | 6 |
Author
| De Ayala, R. J. | 1 |
| DeMars, Christine E. | 1 |
| Dodd, Barbara G. | 1 |
| Jin, Kuan-Yu | 1 |
| Kalinowski, Steven T. | 1 |
| Smith, Richard M. | 1 |
| Wang, Wen-Chung | 1 |
Publication Type
| Journal Articles | 6 |
| Reports - Evaluative | 3 |
| Reports - Research | 3 |
| Speeches/Meeting Papers | 1 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Kalinowski, Steven T. – Educational and Psychological Measurement, 2019
Item response theory (IRT) is a statistical paradigm for developing educational tests and assessing students. IRT, however, currently lacks an established graphical method for examining model fit for the three-parameter logistic model, the most flexible and popular IRT model in educational testing. A method is presented here to do this. The graph,…
Descriptors: Item Response Theory, Educational Assessment, Goodness of Fit, Probability
Jin, Kuan-Yu; Wang, Wen-Chung – Educational and Psychological Measurement, 2014
Extreme response style (ERS) is a systematic tendency for a person to endorse extreme options (e.g., strongly disagree, strongly agree) on Likert-type or rating-scale items. In this study, we develop a new class of item response theory (IRT) models to account for ERS so that the target latent trait is free from the response style and the tendency…
Descriptors: Item Response Theory, Research Methodology, Bayesian Statistics, Response Style (Tests)
DeMars, Christine E. – Educational and Psychological Measurement, 2005
Type I error rates for PARSCALE's fit statistic were examined. Data were generated to fit the partial credit or graded response model, with test lengths of 10 or 20 items. The ability distribution was simulated to be either normal or uniform. Type I error rates were inflated for the shorter test length and, for the graded-response model, also for…
Descriptors: Test Length, Item Response Theory, Psychometrics, Error of Measurement
Peer reviewedSmith, Richard M. – Educational and Psychological Measurement, 1991
This study reports results of an investigation based on simulated data of the distributional properties of the item fit statistics that are commonly used in the Rasch model calibration programs as indices of the fit of responses to individual items to the measurement model. (SLD)
Descriptors: Computer Simulation, Equations (Mathematics), Goodness of Fit, Item Response Theory
Peer reviewedDodd, Barbara G.; And Others – Educational and Psychological Measurement, 1993
Effects of the following variables on performance of computerized adaptive testing (CAT) procedures for the partial credit model (PCM) were studied: (1) stopping rule for terminating CAT; (2) item pool size; and (3) distribution of item difficulties. Implications of findings for CAT systems based on the PCM are discussed. (SLD)
Descriptors: Adaptive Testing, Computer Assisted Testing, Computer Simulation, Difficulty Level
Peer reviewedDe Ayala, R. J. – Educational and Psychological Measurement, 1992
Effects of dimensionality on ability estimation of an adaptive test were examined using generated data in Bayesian computerized adaptive testing (CAT) simulations. Generally, increasing interdimensional difficulty association produced a slight decrease in test length and an increase in accuracy of ability estimation as assessed by root mean square…
Descriptors: Adaptive Testing, Bayesian Statistics, Computer Assisted Testing, Computer Simulation

Direct link
