Publication Date
In 2025 | 0 |
Since 2024 | 2 |
Since 2021 (last 5 years) | 5 |
Since 2016 (last 10 years) | 17 |
Since 2006 (last 20 years) | 43 |
Descriptor
Source
Journal of Educational and… | 48 |
Author
Sinharay, Sandip | 4 |
Jeon, Minjeong | 3 |
Rabe-Hesketh, Sophia | 3 |
De Boeck, Paul | 2 |
Fox, Jean-Paul | 2 |
Magis, David | 2 |
Oranje, Andreas | 2 |
Ranger, Jochen | 2 |
Rijmen, Frank | 2 |
Van den Noortgate, Wim | 2 |
de la Torre, Jimmy | 2 |
More ▼ |
Publication Type
Journal Articles | 48 |
Reports - Research | 29 |
Reports - Descriptive | 15 |
Reports - Evaluative | 5 |
Education Level
Elementary Education | 6 |
Higher Education | 4 |
Intermediate Grades | 4 |
Elementary Secondary Education | 3 |
Grade 4 | 3 |
Grade 8 | 3 |
Junior High Schools | 3 |
Middle Schools | 3 |
Postsecondary Education | 3 |
Secondary Education | 3 |
Grade 1 | 1 |
More ▼ |
Audience
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of… | 4 |
Trends in International… | 2 |
Program for International… | 1 |
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Justin L. Kern – Journal of Educational and Behavioral Statistics, 2024
Given the frequent presence of slipping and guessing in item responses, models for the inclusion of their effects are highly important. Unfortunately, the most common model for their inclusion, the four-parameter item response theory model, potentially has severe deficiencies related to its possible unidentifiability. With this issue in mind, the…
Descriptors: Item Response Theory, Models, Bayesian Statistics, Generalization
Cross-Classified Item Response Theory Modeling with an Application to Student Evaluation of Teaching
Sijia Huang; Li Cai – Journal of Educational and Behavioral Statistics, 2024
The cross-classified data structure is ubiquitous in education, psychology, and health outcome sciences. In these areas, assessment instruments that are made up of multiple items are frequently used to measure latent constructs. The presence of both the cross-classified structure and multivariate categorical outcomes leads to the so-called…
Descriptors: Classification, Data Collection, Data Analysis, Item Response Theory
Sinharay, Sandip – Journal of Educational and Behavioral Statistics, 2022
Takers of educational tests often receive proficiency levels instead of or in addition to scaled scores. For example, proficiency levels are reported for the Advanced Placement (AP®) and U.S. Medical Licensing examinations. Technical difficulties and other unforeseen events occasionally lead to missing item scores and hence to incomplete data on…
Descriptors: Computation, Data Analysis, Educational Testing, Accuracy
Andersson, Björn; Xin, Tao – Journal of Educational and Behavioral Statistics, 2021
The estimation of high-dimensional latent regression item response theory (IRT) models is difficult because of the need to approximate integrals in the likelihood function. Proposed solutions in the literature include using stochastic approximations, adaptive quadrature, and Laplace approximations. We propose using a second-order Laplace…
Descriptors: Item Response Theory, Computation, Regression (Statistics), Statistical Bias
Ramsay, James; Wiberg, Marie; Li, Juan – Journal of Educational and Behavioral Statistics, 2020
Ramsay and Wiberg used a new version of item response theory that represents test performance over nonnegative closed intervals such as [0, 100] or [0, n] and demonstrated that optimal scoring of binary test data yielded substantial improvements in point-wise root-mean-squared error and bias over number right or sum scoring. We extend these…
Descriptors: Scoring, Weighted Scores, Item Response Theory, Intervals
Wang, Shiyu; Xiao, Houping; Cohen, Allan – Journal of Educational and Behavioral Statistics, 2021
An adaptive weight estimation approach is proposed to provide robust latent ability estimation in computerized adaptive testing (CAT) with response revision. This approach assigns different weights to each distinct response to the same item when response revision is allowed in CAT. Two types of weight estimation procedures, nonfunctional and…
Descriptors: Computer Assisted Testing, Adaptive Testing, Computation, Robustness (Statistics)
Köhler, Carmen; Robitzsch, Alexander; Hartig, Johannes – Journal of Educational and Behavioral Statistics, 2020
Testing whether items fit the assumptions of an item response theory model is an important step in evaluating a test. In the literature, numerous item fit statistics exist, many of which show severe limitations. The current study investigates the root mean squared deviation (RMSD) item fit statistic, which is used for evaluating item fit in…
Descriptors: Test Items, Goodness of Fit, Statistics, Bias
Barrett, Michelle D.; van der Linden, Wim J. – Journal of Educational and Behavioral Statistics, 2019
Parameter linking in item response theory is generally necessary to adjust for differences between the true values for the same item and ability parameters due to the use of different identifiability restrictions in different calibrations. The research reported in this article explores a precision-weighted (PW) approach to the problem of…
Descriptors: Item Response Theory, Computation, Error of Measurement, Test Items
van der Linden, Wim J.; Ren, Hao – Journal of Educational and Behavioral Statistics, 2020
The Bayesian way of accounting for the effects of error in the ability and item parameters in adaptive testing is through the joint posterior distribution of all parameters. An optimized Markov chain Monte Carlo algorithm for adaptive testing is presented, which samples this distribution in real time to score the examinee's ability and optimally…
Descriptors: Bayesian Statistics, Adaptive Testing, Error of Measurement, Markov Processes
Jewsbury, Paul A.; van Rijn, Peter W. – Journal of Educational and Behavioral Statistics, 2020
In large-scale educational assessment data consistent with a simple-structure multidimensional item response theory (MIRT) model, where every item measures only one latent variable, separate unidimensional item response theory (UIRT) models for each latent variable are often calibrated for practical reasons. While this approach can be valid for…
Descriptors: Item Response Theory, Computation, Test Items, Adaptive Testing
Choi, Jinnie – Journal of Educational and Behavioral Statistics, 2017
This article reviews PROC IRT, which was added to Statistical Analysis Software in 2014. We provide an introductory overview of a free version of SAS, describe what PROC IRT offers for item response theory (IRT) analysis and how one can use PROC IRT, and discuss how other SAS macros and procedures may compensate the IRT functionalities of PROC IRT.
Descriptors: Item Response Theory, Computer Software, Statistical Analysis, Computation
Choe, Edison M.; Kern, Justin L.; Chang, Hua-Hua – Journal of Educational and Behavioral Statistics, 2018
Despite common operationalization, measurement efficiency of computerized adaptive testing should not only be assessed in terms of the number of items administered but also the time it takes to complete the test. To this end, a recent study introduced a novel item selection criterion that maximizes Fisher information per unit of expected response…
Descriptors: Computer Assisted Testing, Reaction Time, Item Response Theory, Test Items
Oranje, Andreas; Kolstad, Andrew – Journal of Educational and Behavioral Statistics, 2019
The design and psychometric methodology of the National Assessment of Educational Progress (NAEP) is constantly evolving to meet the changing interests and demands stemming from a rapidly shifting educational landscape. NAEP has been built on strong research foundations that include conducting extensive evaluations and comparisons before new…
Descriptors: National Competency Tests, Psychometrics, Statistical Analysis, Computation
Ranger, Jochen; Kuhn, Jörg-Tobias – Journal of Educational and Behavioral Statistics, 2018
Diffusion-based item response theory models for responses and response times in tests have attracted increased attention recently in psychometrics. Analyzing response time data, however, is delicate as response times are often contaminated by unusual observations. This can have serious effects on the validity of statistical inference. In this…
Descriptors: Item Response Theory, Computation, Robustness (Statistics), Reaction Time
Magis, David – Journal of Educational and Behavioral Statistics, 2015
The purpose of this note is to study the equivalence of observed and expected (Fisher) information functions with polytomous item response theory (IRT) models. It is established that observed and expected information functions are equivalent for the class of divide-by-total models (including partial credit, generalized partial credit, rating…
Descriptors: Item Response Theory, Models, Statistics, Computation