NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 12 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Diao, Qi; van der Linden, Wim J. – Applied Psychological Measurement, 2013
Automated test assembly uses the methodology of mixed integer programming to select an optimal set of items from an item bank. Automated test-form generation uses the same methodology to optimally order the items and format the test form. From an optimization point of view, production of fully formatted test forms directly from the item pool using…
Descriptors: Automation, Test Construction, Test Format, Item Banks
Peer reviewed Peer reviewed
Direct linkDirect link
de la Torre, Jimmy; Song, Hao; Hong, Yuan – Applied Psychological Measurement, 2011
Lack of sufficient reliability is the primary impediment for generating and reporting subtest scores. Several current methods of subscore estimation do so either by incorporating the correlational structure among the subtest abilities or by using the examinee's performance on the overall test. This article conducted a systematic comparison of four…
Descriptors: Item Response Theory, Scoring, Methods, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
van der Linden, Wim J.; Wiberg, Marie – Applied Psychological Measurement, 2010
For traditional methods of observed-score equating with anchor-test designs, such as chain and poststratification equating, it is difficult to satisfy the criteria of equity and population invariance. Their equatings are therefore likely to be biased. The bias in these methods was evaluated against a simple local equating method in which the…
Descriptors: Methods, Equated Scores, Test Items, Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Chen, Shu-Ying – Applied Psychological Measurement, 2010
To date, exposure control procedures that are designed to control test overlap in computerized adaptive tests (CATs) are based on the assumption of item sharing between pairs of examinees. However, in practice, examinees may obtain test information from more than one previous test taker. This larger scope of information sharing needs to be…
Descriptors: Computer Assisted Testing, Adaptive Testing, Methods, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Paek, Insu; Guo, Hongwen – Applied Psychological Measurement, 2011
This study examined how much improvement was attainable with respect to accuracy of differential item functioning (DIF) measures and DIF detection rates in the Mantel-Haenszel procedure when employing focal and reference groups with notably unbalanced sample sizes where the focal group has a fixed small sample which does not satisfy the minimum…
Descriptors: Test Bias, Accuracy, Reference Groups, Investigations
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Wen-Chung; Shih, Ching-Lin – Applied Psychological Measurement, 2010
Three multiple indicators-multiple causes (MIMIC) methods, namely, the standard MIMIC method (M-ST), the MIMIC method with scale purification (M-SP), and the MIMIC method with a pure anchor (M-PA), were developed to assess differential item functioning (DIF) in polytomous items. In a series of simulations, it appeared that all three methods…
Descriptors: Methods, Test Bias, Test Items, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Woods, Carol M. – Applied Psychological Measurement, 2009
Differential item functioning (DIF) occurs when an item on a test or questionnaire has different measurement properties for one group of people versus another, irrespective of mean differences on the construct. There are many methods available for DIF assessment. The present article is focused on indices of partial association. A family of average…
Descriptors: Test Bias, Measurement, Correlation, Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Levy, Roy; Mislevy, Robert J.; Sinharay, Sandip – Applied Psychological Measurement, 2009
If data exhibit multidimensionality, key conditional independence assumptions of unidimensional models do not hold. The current work pursues posterior predictive model checking, a flexible family of model-checking procedures, as a tool for criticizing models due to unaccounted for dimensions in the context of item response theory. Factors…
Descriptors: Item Response Theory, Models, Methods, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Shih, Ching-Lin; Wang, Wen-Chung – Applied Psychological Measurement, 2009
The multiple indicators, multiple causes (MIMIC) method with a pure short anchor was proposed to detect differential item functioning (DIF). A simulation study showed that the MIMIC method with an anchor of 1, 2, 4, or 10 DIF-free items yielded a well-controlled Type I error rate even when such tests contained as many as 40% DIF items. In general,…
Descriptors: Test Bias, Simulation, Methods, Factor Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Riley, Barth B.; Dennis, Michael L.; Conrad, Kendon J. – Applied Psychological Measurement, 2010
This simulation study sought to compare four different computerized adaptive testing (CAT) content-balancing procedures designed for use in a multidimensional assessment with respect to measurement precision, symptom severity classification, validity of clinical diagnostic recommendations, and sensitivity to atypical responding. The four…
Descriptors: Simulation, Computer Assisted Testing, Adaptive Testing, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Li, Feiming; Cohen, Allan S.; Kim, Seock-Ho; Cho, Sun-Joo – Applied Psychological Measurement, 2009
This study examines model selection indices for use with dichotomous mixture item response theory (IRT) models. Five indices are considered: Akaike's information coefficient (AIC), Bayesian information coefficient (BIC), deviance information coefficient (DIC), pseudo-Bayes factor (PsBF), and posterior predictive model checks (PPMC). The five…
Descriptors: Item Response Theory, Models, Selection, Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Ip, Edward H.; Smits, Dirk J. M.; De Boeck, Paul – Applied Psychological Measurement, 2009
The article proposes a family of item-response models that allow the separate and independent specification of three orthogonal components: item attribute, person covariate, and local item dependence. Special interest lies in extending the linear logistic test model, which is commonly used to measure item attributes, to tests with embedded item…
Descriptors: Item Response Theory, Models, Computation, Methods