Publication Date
| In 2026 | 0 |
| Since 2025 | 1 |
| Since 2022 (last 5 years) | 4 |
| Since 2017 (last 10 years) | 42 |
| Since 2007 (last 20 years) | 110 |
Descriptor
| Computation | 117 |
| Statistical Analysis | 117 |
| Item Response Theory | 98 |
| Models | 50 |
| Test Items | 35 |
| Comparative Analysis | 26 |
| Simulation | 25 |
| Bayesian Statistics | 19 |
| Maximum Likelihood Statistics | 19 |
| Scores | 18 |
| Error of Measurement | 17 |
| More ▼ | |
Source
Author
| Cho, Sun-Joo | 5 |
| Sinharay, Sandip | 3 |
| Wang, Wen-Chung | 3 |
| Bottge, Brian | 2 |
| Bulut, Okan | 2 |
| Cai, Li | 2 |
| Castellano, Katherine E. | 2 |
| Chuang, Chi-ching | 2 |
| Cohen, Allan S. | 2 |
| Dimitrov, Dimiter M. | 2 |
| Dorans, Neil J. | 2 |
| More ▼ | |
Publication Type
| Journal Articles | 101 |
| Reports - Research | 87 |
| Reports - Evaluative | 16 |
| Reports - Descriptive | 10 |
| Speeches/Meeting Papers | 5 |
| Dissertations/Theses -… | 4 |
| Opinion Papers | 2 |
| Tests/Questionnaires | 2 |
| Collected Works - Proceedings | 1 |
Education Level
Audience
| Practitioners | 1 |
| Students | 1 |
| Teachers | 1 |
Location
| Germany | 2 |
| Massachusetts | 2 |
| Netherlands | 2 |
| Taiwan | 2 |
| Australia | 1 |
| Botswana | 1 |
| Brazil | 1 |
| Canada | 1 |
| Chile | 1 |
| Czech Republic | 1 |
| Ghana | 1 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Jianbin Fu; TsungHan Ho; Xuan Tan – Practical Assessment, Research & Evaluation, 2025
Item parameter estimation using an item response theory (IRT) model with fixed ability estimates is useful in equating with small samples on anchor items. The current study explores the impact of three ability estimation methods (weighted likelihood estimation [WLE], maximum a posteriori [MAP], and posterior ability distribution estimation [PST])…
Descriptors: Item Response Theory, Test Items, Computation, Equated Scores
Nianbo Dong; Benjamin Kelcey; Jessaca Spybrook; Yanli Xie; Dung Pham; Peilin Qiu; Ning Sui – Grantee Submission, 2024
Multisite trials that randomize individuals (e.g., students) within sites (e.g., schools) or clusters (e.g., teachers/classrooms) within sites (e.g., schools) are commonly used for program evaluation because they provide opportunities to learn about treatment effects as well as their heterogeneity across sites and subgroups (defined by moderating…
Descriptors: Statistical Analysis, Randomized Controlled Trials, Educational Research, Effect Size
Rachatasumrit, Napol; Koedinger, Kenneth R. – International Educational Data Mining Society, 2021
Student modeling is useful in educational research and technology development due to a capability to estimate latent student attributes. Widely used approaches, such as the Additive Factors Model (AFM), have shown satisfactory results, but they can only handle binary outcomes, which may yield potential information loss. In this work, we propose a…
Descriptors: Models, Student Characteristics, Feedback (Response), Error Correction
Zhou, Sherry; Huggins-Manley, Anne Corinne – Educational and Psychological Measurement, 2020
The semi-generalized partial credit model (Semi-GPCM) has been proposed as a unidimensional modeling method for handling not applicable scale responses and neutral scale responses, and it has been suggested that the model may be of use in handling missing data in scale items. The purpose of this study is to evaluate the ability of the…
Descriptors: Models, Statistical Analysis, Response Style (Tests), Test Items
Raykov, Tenko; Marcoulides, George A.; Pusic, Martin – Measurement: Interdisciplinary Research and Perspectives, 2021
An interval estimation procedure is discussed that can be used to evaluate the probability of a particular response for a binary or binary scored item at a pre-specified point along an underlying latent continuum. The item is assumed to: (a) be part of a unidimensional multi-component measuring instrument that may contain also polytomous items,…
Descriptors: Item Response Theory, Computation, Probability, Test Items
Robitzsch, Alexander – Journal of Intelligence, 2020
The last series of Raven's standard progressive matrices (SPM-LS) test was studied with respect to its psychometric properties in a series of recent papers. In this paper, the SPM-LS dataset is analyzed with regularized latent class models (RLCMs). For dichotomous item response data, an alternative estimation approach based on fused regularization…
Descriptors: Statistical Analysis, Classification, Intelligence Tests, Test Items
Alahmadi, Sarah; Jones, Andrew T.; Barry, Carol L.; Ibáñez, Beatriz – Applied Measurement in Education, 2023
Rasch common-item equating is often used in high-stakes testing to maintain equivalent passing standards across test administrations. If unaddressed, item parameter drift poses a major threat to the accuracy of Rasch common-item equating. We compared the performance of well-established and newly developed drift detection methods in small and large…
Descriptors: Equated Scores, Item Response Theory, Sample Size, Test Items
Tang, Xiaodan; Karabatsos, George; Chen, Haiqin – Applied Measurement in Education, 2020
In applications of item response theory (IRT) models, it is known that empirical violations of the local independence (LI) assumption can significantly bias parameter estimates. To address this issue, we propose a threshold-autoregressive item response theory (TAR-IRT) model that additionally accounts for order dependence among the item responses…
Descriptors: Item Response Theory, Test Items, Models, Computation
Johnson, Michael Lee; Sedory, Stephen A.; Singh, Sarjinder – Sociological Methods & Research, 2019
In this article, alternative randomized response models are proposed, which make use of sum of quantitative scores generated from two decks of cards being used in a survey. The proposed methods are compared to the Odumade and Singh and Singh and Grewal models through a simulation study. It is shown that the modified methods can be used more…
Descriptors: Scores, Surveys, Responses, Models
Lozano, José H.; Revuelta, Javier – Applied Measurement in Education, 2021
The present study proposes a Bayesian approach for estimating and testing the operation-specific learning model, a variant of the linear logistic test model that allows for the measurement of the learning that occurs during a test as a result of the repeated use of the operations involved in the items. The advantages of using a Bayesian framework…
Descriptors: Bayesian Statistics, Computation, Learning, Testing
Diaz, Emily; Brooks, Gordon; Johanson, George – International Journal of Assessment Tools in Education, 2021
This Monte Carlo study assessed Type I error in differential item functioning analyses using Lord's chi-square (LC), Likelihood Ratio Test (LRT), and Mantel-Haenszel (MH) procedure. Two research interests were investigated: item response theory (IRT) model specification in LC and the LRT and continuity correction in the MH procedure. This study…
Descriptors: Test Bias, Item Response Theory, Statistical Analysis, Comparative Analysis
Guo, Hongwen; Dorans, Neil J. – ETS Research Report Series, 2019
We derive formulas for the differential item functioning (DIF) measures that two routinely used DIF statistics are designed to estimate. The DIF measures that match on observed scores are compared to DIF measures based on an unobserved ability (theta or true score) for items that are described by either the one-parameter logistic (1PL) or…
Descriptors: Scores, Test Bias, Statistical Analysis, Item Response Theory
Joshua B. Gilbert – Annenberg Institute for School Reform at Brown University, 2022
This simulation study examines the characteristics of the Explanatory Item Response Model (EIRM) when estimating treatment effects when compared to classical test theory (CTT) sum and mean scores and item response theory (IRT)-based theta scores. Results show that the EIRM and IRT theta scores provide generally equivalent bias and false positive…
Descriptors: Item Response Theory, Models, Test Theory, Computation
Xiao, Jiaying; Bulut, Okan – Educational and Psychological Measurement, 2020
Large amounts of missing data could distort item parameter estimation and lead to biased ability estimates in educational assessments. Therefore, missing responses should be handled properly before estimating any parameters. In this study, two Monte Carlo simulation studies were conducted to compare the performance of four methods in handling…
Descriptors: Data, Computation, Ability, Maximum Likelihood Statistics
Cho, April E.; Wang, Chun; Zhang, Xue; Xu, Gongjun – Grantee Submission, 2020
Multidimensional Item Response Theory (MIRT) is widely used in assessment and evaluation of educational and psychological tests. It models the individual response patterns by specifying functional relationship between individuals' multiple latent traits and their responses to test items. One major challenge in parameter estimation in MIRT is that…
Descriptors: Item Response Theory, Mathematics, Statistical Inference, Maximum Likelihood Statistics

Peer reviewed
Direct link
