Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 96 |
Descriptor
Computation | 107 |
Item Response Theory | 70 |
Models | 45 |
Simulation | 38 |
Test Items | 25 |
Monte Carlo Methods | 21 |
Bayesian Statistics | 20 |
Evaluation Methods | 20 |
Statistical Analysis | 20 |
Computer Software | 19 |
Error of Measurement | 19 |
More ▼ |
Source
Applied Psychological… | 107 |
Author
de la Torre, Jimmy | 5 |
Wang, Wen-Chung | 4 |
Woods, Carol M. | 4 |
Brennan, Robert L. | 3 |
Cho, Sun-Joo | 3 |
Cohen, Allan S. | 3 |
De Boeck, Paul | 3 |
Kim, Seock-Ho | 3 |
Kreiner, Svend | 3 |
Magis, David | 3 |
Raiche, Gilles | 3 |
More ▼ |
Publication Type
Journal Articles | 107 |
Reports - Research | 54 |
Reports - Evaluative | 29 |
Reports - Descriptive | 24 |
Education Level
Audience
Laws, Policies, & Programs
Assessments and Surveys
California Achievement Tests | 2 |
Eysenck Personality Inventory | 2 |
ACT Assessment | 1 |
Armed Services Vocational… | 1 |
Florida Comprehensive… | 1 |
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Lei, Pui-Wa; Li, Hongli – Applied Psychological Measurement, 2013
Minimum sample sizes of about 200 to 250 per group are often recommended for differential item functioning (DIF) analyses. However, there are times when sample sizes for one or both groups of interest are smaller than 200 due to practical constraints. This study attempts to examine the performance of Simultaneous Item Bias Test (SIBTEST),…
Descriptors: Sample Size, Test Bias, Computation, Accuracy
Jingchen Liu; Gongjun Xu; Zhiliang Ying – Applied Psychological Measurement, 2012
The recent surge of interests in cognitive assessment has led to developments of novel statistical models for diagnostic classification. Central to many such models is the well-known "Q"-matrix, which specifies the item-attribute relationships. This article proposes a data-driven approach to identification of the "Q"-matrix and…
Descriptors: Matrices, Computation, Statistical Analysis, Models
Chen, Jinsong; de la Torre, Jimmy – Applied Psychological Measurement, 2013
Polytomous attributes, particularly those defined as part of the test development process, can provide additional diagnostic information. The present research proposes the polytomous generalized deterministic inputs, noisy, "and" gate (pG-DINA) model to accommodate such attributes. The pG-DINA model allows input from substantive experts…
Descriptors: Models, Cognitive Tests, Diagnostic Tests, Computation
Li, Ying; Lissitz, Robert W. – Applied Psychological Measurement, 2012
To address the lack of attention to construct shift in item response theory (IRT) vertical scaling, a multigroup, bifactor model was proposed to model the common dimension for all grades and the grade-specific dimensions. Bifactor model estimation accuracy was evaluated through a simulation study with manipulated factors of percentage of common…
Descriptors: Item Response Theory, Scaling, Models, Computation
Lathrop, Quinn N.; Cheng, Ying – Applied Psychological Measurement, 2013
Within the framework of item response theory (IRT), there are two recent lines of work on the estimation of classification accuracy (CA) rate. One approach estimates CA when decisions are made based on total sum scores, the other based on latent trait estimates. The former is referred to as the Lee approach, and the latter, the Rudner approach,…
Descriptors: Item Response Theory, Accuracy, Classification, Computation
Lei, Pui-Wa; Zhao, Yu – Applied Psychological Measurement, 2012
Vertical scaling is necessary to facilitate comparison of scores from test forms of different difficulty levels. It is widely used to enable the tracking of student growth in academic performance over time. Most previous studies on vertical scaling methods assume relatively long tests and large samples. Little is known about their performance when…
Descriptors: Scaling, Item Response Theory, Test Length, Sample Size
Nandakumar, Ratna; Hotchkiss, Lawrence – Applied Psychological Measurement, 2012
The PROC NLMIXED procedure in Statistical Analysis System can be used to estimate parameters of item response theory (IRT) models. The data for this procedure are set up in a particular format called the "long format." The long format takes a substantial amount of time to execute the program. This article describes a format called the "wide…
Descriptors: Item Response Theory, Models, Statistical Analysis, Computer Software
Kalender, Ilker – Applied Psychological Measurement, 2012
catcher is a software program designed to compute the [omega] index, a common statistical index for the identification of collusions (cheating) among examinees taking an educational or psychological test. It requires (a) responses and (b) ability estimations of individuals, and (c) item parameters to make computations and outputs the results of…
Descriptors: Computer Software, Computation, Statistical Analysis, Cheating
Culpepper, Steven Andrew – Applied Psychological Measurement, 2013
A classic topic in the fields of psychometrics and measurement has been the impact of the number of scale categories on test score reliability. This study builds on previous research by further articulating the relationship between item response theory (IRT) and classical test theory (CTT). Equations are presented for comparing the reliability and…
Descriptors: Item Response Theory, Reliability, Scores, Error of Measurement
Paek, Insu; Han, Kyung T. – Applied Psychological Measurement, 2013
This article reviews a new item response theory (IRT) model estimation program, IRTPRO 2.1, for Windows that is capable of unidimensional and multidimensional IRT model estimation for existing and user-specified constrained IRT models for dichotomously and polytomously scored item response data. (Contains 1 figure and 2 notes.)
Descriptors: Item Response Theory, Computer Software, Computation, Patients
Liu, Yang; Thissen, David – Applied Psychological Measurement, 2012
Local dependence (LD) refers to the violation of the local independence assumption of most item response models. Statistics that indicate LD between a pair of items on a test or questionnaire that is being fitted with an item response model can play a useful diagnostic role in applications of item response theory. In this article, a new score test…
Descriptors: Item Response Theory, Statistical Analysis, Models, Identification
Dai, Yunyun – Applied Psychological Measurement, 2013
Mixtures of item response theory (IRT) models have been proposed as a technique to explore response patterns in test data related to cognitive strategies, instructional sensitivity, and differential item functioning (DIF). Estimation proves challenging due to difficulties in identification and questions of effect size needed to recover underlying…
Descriptors: Item Response Theory, Test Bias, Computation, Bayesian Statistics
De Boeck, Paul; Cho, Sun-Joo; Wilson, Mark – Applied Psychological Measurement, 2011
The models used in this article are secondary dimension mixture models with the potential to explain differential item functioning (DIF) between latent classes, called latent DIF. The focus is on models with a secondary dimension that is at the same time specific to the DIF latent class and linked to an item property. A description of the models…
Descriptors: Test Bias, Models, Statistical Analysis, Computation
Andrich, David; Humphry, Stephen M.; Marais, Ida – Applied Psychological Measurement, 2012
Models of modern test theory imply statistical independence among responses, generally referred to as "local independence." One violation of local independence occurs when the response to one item governs the response to a subsequent item. Expanding on a formulation of this kind of violation as a process in the dichotomous Rasch model,…
Descriptors: Test Theory, Models, Item Response Theory, Evidence
Ip, Edward Hak-Sing; Chen, Shyh-Huei – Applied Psychological Measurement, 2012
The problem of fitting unidimensional item-response models to potentially multidimensional data has been extensively studied. The focus of this article is on response data that contains a major dimension of interest but that may also contain minor nuisance dimensions. Because fitting a unidimensional model to multidimensional data results in…
Descriptors: Measurement, Item Response Theory, Scores, Computation