NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 26 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Zwick, Rebecca; Ye, Lei; Isham, Steven – Journal of Educational Measurement, 2018
In typical differential item functioning (DIF) assessments, an item's DIF status is not influenced by its status in previous test administrations. An item that has shown DIF at multiple administrations may be treated the same way as an item that has shown DIF in only the most recent administration. Therefore, much useful information about the…
Descriptors: Test Bias, Testing, Test Items, Bayesian Statistics
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Zwick, Rebecca; Ye, Lei; Isham, Steven – ETS Research Report Series, 2013
Differential item functioning (DIF) analysis is a key component in the evaluation of the fairness and validity of educational tests. Although it is often assumed that refinement of the matching criterion always provides more accurate DIF results, the actual situation proves to be more complex. To explore the effectiveness of refinement, we…
Descriptors: Test Bias, Statistical Analysis, Simulation, Educational Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Zwick, Rebecca; Ye, Lei; Isham, Steven – Journal of Educational and Behavioral Statistics, 2012
This study demonstrates how the stability of Mantel-Haenszel (MH) DIF (differential item functioning) methods can be improved by integrating information across multiple test administrations using Bayesian updating (BU). The authors conducted a simulation that showed that this approach, which is based on earlier work by Zwick, Thayer, and Lewis,…
Descriptors: Test Bias, Computation, Statistical Analysis, Bayesian Statistics
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Zwick, Rebecca – ETS Research Report Series, 2012
Differential item functioning (DIF) analysis is a key component in the evaluation of the fairness and validity of educational tests. The goal of this project was to review the status of ETS DIF analysis procedures, focusing on three aspects: (a) the nature and stringency of the statistical rules used to flag items, (b) the minimum sample size…
Descriptors: Test Bias, Sample Size, Bayesian Statistics, Evaluation Methods
Peer reviewed Peer reviewed
Zwick, Rebecca; Thayer, Dorothy T.; Lewis, Charles – Journal of Educational Measurement, 1999
Developed an empirical Bayes enhancement to Mantel-Haenszel (MH) analysis of differential item functioning (DIF) in which it is assumed that the MH statistics are normally distributed and that the prior distribution of underlying DIF parameters is also normal. (Author/SLD)
Descriptors: Bayesian Statistics, Item Bias, Statistical Distributions, Test Items
Peer reviewed Peer reviewed
Zwick, Rebecca; Thayer, Dorothy; Lewis, Charles – Journal of Educational and Behavioral Statistics, 2000
Studied a method for flagging differential item functioning (DIF) based on loss functions. Builds on earlier research that led to the development of an empirical Bayes enhancement to the Mantel-Haenszel DIF analysis. Tested the method through simulation and found its performance better than some commonly used DIF classification systems. (SLD)
Descriptors: Bayesian Statistics, Identification, Item Bias, Simulation
Peer reviewed Peer reviewed
Zwick, Rebecca; Thayer, Dorothy T.; Mazzeo, John – Applied Measurement in Education, 1997
Differential item functioning (DIF) assessment procedures for items with more than two ordered score categories, referred to as polytomous items, were evaluated. Three descriptive statistics (standardized mean difference and two procedures based on the SIBTEST computer program) and five inferential procedures were used. Conditions under which the…
Descriptors: Item Bias, Research Methodology, Statistical Inference, Test Construction
Peer reviewed Peer reviewed
Zwick, Rebecca – Educational and Psychological Measurement, 1997
Recent simulations have shown that, for a given sample size, the Mantel-Haenszel (MH) variances tend to be larger when items are administered to randomly selected examinees than when they are administered adaptively. Results suggest that adaptive testing may lead to more efficient application of MH differential item functioning analyses. (SLD)
Descriptors: Adaptive Testing, Item Bias, Sample Size, Simulation
Peer reviewed Peer reviewed
Rudas, Tamas; Zwick, Rebecca – Journal of Educational and Behavioral Statistics, 1997
The mixture index of fit (T. Rudas et al, 1994) is used to estimate the fraction of a population for which differential item functioning (DIF) occurs, and this approach is compared to the Mantel Haenszel test of DIF. The proposed noniterative procedure provides information about data portions contributing to DIF. (SLD)
Descriptors: Comparative Analysis, Estimation (Mathematics), Item Bias, Maximum Likelihood Statistics
Peer reviewed Peer reviewed
Zwick, Rebecca; Thayer, Dorothy T. – Applied Psychological Measurement, 2002
Used a simulation to investigate the applicability to computerized adaptive test data of a differential item functioning (DIF) analysis method. Results show the performance of this empirical Bayes enhancement of the Mantel Haenszel DIF analysis method to be quite promising. (SLD)
Descriptors: Adaptive Testing, Bayesian Statistics, Computer Assisted Testing, Item Bias
Zwick, Rebecca – 1994
The Mantel Haenszel (MH; 1959) approach of Holland and Thayer (1988) is a well-established method for assessing differential item functioning (DIF). The formula for the variance of the MH DIF statistic is based on work by Phillips and Holland (1987) and Robins, Breslow, and Greenland (1986). Recent simulation studies showed that the MH variances…
Descriptors: Adaptive Testing, Evaluation Methods, Item Bias, Measurement Techniques
Peer reviewed Peer reviewed
Zwick, Rebecca; Thayer, Dorothy T. – Journal of Educational and Behavioral Statistics, 1996
Two possible standard error formulas for the polytomous differential item functioning index proposed by N. J. Dorans and A. P. Schmitt (1991) were derived. These standard errors, and associated hypothesis-testing procedures, were evaluated through simulated data. The standard error that performed better is based on N. Mantel's (1963)…
Descriptors: Error of Measurement, Evaluation Methods, Hypothesis Testing, Item Bias
Peer reviewed Peer reviewed
Zwick, Rebecca – Change, 2001
Considers the guide on high-stakes testing issued by the federal Office for Civil Rights, including the controversy which ensued upon release of the first draft, changes in the subsequent version, and the issue of differences in educational achievement among ethnic and racial groups of which differences in standardized test scores may be…
Descriptors: Academic Achievement, Guides, High Stakes Tests, Minority Groups
Zwick, Rebecca; Thayer, Dorothy T. – 2003
This study investigated the applicability to computerized adaptive testing (CAT) data of a differential item functioning (DIF) analysis that involves an empirical Bayes (EB) enhancement of the popular Mantel Haenszel (MH) DIF analysis method. The computerized Law School Admission Test (LSAT) assumed for this study was similar to that currently…
Descriptors: Adaptive Testing, Bayesian Statistics, College Entrance Examinations, Computer Assisted Testing
Peer reviewed Peer reviewed
Zwick, Rebecca; And Others – Journal of Educational Measurement, 1995
In a simulation study of ability and estimation of differential item functioning (DIF) in computerized adaptive tests, Rasch-based DIF statistics were highly correlated with generating DIF, but DIF statistics tended to be slightly smaller than in the three-parameter logistic model analyses. (SLD)
Descriptors: Ability, Adaptive Testing, Computer Assisted Testing, Computer Simulation
Previous Page | Next Page ยป
Pages: 1  |  2