Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 4 |
Descriptor
Source
Journal of Educational… | 4 |
Journal of Educational and… | 4 |
Applied Psychological… | 2 |
ETS Research Report Series | 2 |
Applied Measurement in… | 1 |
Change | 1 |
Educational and Psychological… | 1 |
Journal of Educational… | 1 |
Author
Zwick, Rebecca | 26 |
Thayer, Dorothy T. | 6 |
Isham, Steven | 3 |
Ye, Lei | 3 |
Lewis, Charles | 2 |
Rudas, Tamas | 2 |
Ercikan, Kadriye | 1 |
Mazzeo, John | 1 |
Thayer, Dorothy | 1 |
Publication Type
Reports - Evaluative | 18 |
Journal Articles | 16 |
Reports - Research | 7 |
Speeches/Meeting Papers | 3 |
Reports - Descriptive | 1 |
Education Level
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
Advanced Placement… | 1 |
Law School Admission Test | 1 |
National Assessment of… | 1 |
What Works Clearinghouse Rating
Zwick, Rebecca; Ye, Lei; Isham, Steven – Journal of Educational Measurement, 2018
In typical differential item functioning (DIF) assessments, an item's DIF status is not influenced by its status in previous test administrations. An item that has shown DIF at multiple administrations may be treated the same way as an item that has shown DIF in only the most recent administration. Therefore, much useful information about the…
Descriptors: Test Bias, Testing, Test Items, Bayesian Statistics
Zwick, Rebecca; Ye, Lei; Isham, Steven – ETS Research Report Series, 2013
Differential item functioning (DIF) analysis is a key component in the evaluation of the fairness and validity of educational tests. Although it is often assumed that refinement of the matching criterion always provides more accurate DIF results, the actual situation proves to be more complex. To explore the effectiveness of refinement, we…
Descriptors: Test Bias, Statistical Analysis, Simulation, Educational Testing
Zwick, Rebecca; Ye, Lei; Isham, Steven – Journal of Educational and Behavioral Statistics, 2012
This study demonstrates how the stability of Mantel-Haenszel (MH) DIF (differential item functioning) methods can be improved by integrating information across multiple test administrations using Bayesian updating (BU). The authors conducted a simulation that showed that this approach, which is based on earlier work by Zwick, Thayer, and Lewis,…
Descriptors: Test Bias, Computation, Statistical Analysis, Bayesian Statistics
Zwick, Rebecca – ETS Research Report Series, 2012
Differential item functioning (DIF) analysis is a key component in the evaluation of the fairness and validity of educational tests. The goal of this project was to review the status of ETS DIF analysis procedures, focusing on three aspects: (a) the nature and stringency of the statistical rules used to flag items, (b) the minimum sample size…
Descriptors: Test Bias, Sample Size, Bayesian Statistics, Evaluation Methods

Zwick, Rebecca; Thayer, Dorothy T.; Lewis, Charles – Journal of Educational Measurement, 1999
Developed an empirical Bayes enhancement to Mantel-Haenszel (MH) analysis of differential item functioning (DIF) in which it is assumed that the MH statistics are normally distributed and that the prior distribution of underlying DIF parameters is also normal. (Author/SLD)
Descriptors: Bayesian Statistics, Item Bias, Statistical Distributions, Test Items

Zwick, Rebecca; Thayer, Dorothy; Lewis, Charles – Journal of Educational and Behavioral Statistics, 2000
Studied a method for flagging differential item functioning (DIF) based on loss functions. Builds on earlier research that led to the development of an empirical Bayes enhancement to the Mantel-Haenszel DIF analysis. Tested the method through simulation and found its performance better than some commonly used DIF classification systems. (SLD)
Descriptors: Bayesian Statistics, Identification, Item Bias, Simulation

Zwick, Rebecca; Thayer, Dorothy T.; Mazzeo, John – Applied Measurement in Education, 1997
Differential item functioning (DIF) assessment procedures for items with more than two ordered score categories, referred to as polytomous items, were evaluated. Three descriptive statistics (standardized mean difference and two procedures based on the SIBTEST computer program) and five inferential procedures were used. Conditions under which the…
Descriptors: Item Bias, Research Methodology, Statistical Inference, Test Construction

Zwick, Rebecca – Educational and Psychological Measurement, 1997
Recent simulations have shown that, for a given sample size, the Mantel-Haenszel (MH) variances tend to be larger when items are administered to randomly selected examinees than when they are administered adaptively. Results suggest that adaptive testing may lead to more efficient application of MH differential item functioning analyses. (SLD)
Descriptors: Adaptive Testing, Item Bias, Sample Size, Simulation

Rudas, Tamas; Zwick, Rebecca – Journal of Educational and Behavioral Statistics, 1997
The mixture index of fit (T. Rudas et al, 1994) is used to estimate the fraction of a population for which differential item functioning (DIF) occurs, and this approach is compared to the Mantel Haenszel test of DIF. The proposed noniterative procedure provides information about data portions contributing to DIF. (SLD)
Descriptors: Comparative Analysis, Estimation (Mathematics), Item Bias, Maximum Likelihood Statistics

Zwick, Rebecca; Thayer, Dorothy T. – Applied Psychological Measurement, 2002
Used a simulation to investigate the applicability to computerized adaptive test data of a differential item functioning (DIF) analysis method. Results show the performance of this empirical Bayes enhancement of the Mantel Haenszel DIF analysis method to be quite promising. (SLD)
Descriptors: Adaptive Testing, Bayesian Statistics, Computer Assisted Testing, Item Bias
Zwick, Rebecca – 1994
The Mantel Haenszel (MH; 1959) approach of Holland and Thayer (1988) is a well-established method for assessing differential item functioning (DIF). The formula for the variance of the MH DIF statistic is based on work by Phillips and Holland (1987) and Robins, Breslow, and Greenland (1986). Recent simulation studies showed that the MH variances…
Descriptors: Adaptive Testing, Evaluation Methods, Item Bias, Measurement Techniques

Zwick, Rebecca; Thayer, Dorothy T. – Journal of Educational and Behavioral Statistics, 1996
Two possible standard error formulas for the polytomous differential item functioning index proposed by N. J. Dorans and A. P. Schmitt (1991) were derived. These standard errors, and associated hypothesis-testing procedures, were evaluated through simulated data. The standard error that performed better is based on N. Mantel's (1963)…
Descriptors: Error of Measurement, Evaluation Methods, Hypothesis Testing, Item Bias

Zwick, Rebecca – Change, 2001
Considers the guide on high-stakes testing issued by the federal Office for Civil Rights, including the controversy which ensued upon release of the first draft, changes in the subsequent version, and the issue of differences in educational achievement among ethnic and racial groups of which differences in standardized test scores may be…
Descriptors: Academic Achievement, Guides, High Stakes Tests, Minority Groups
Zwick, Rebecca; Thayer, Dorothy T. – 2003
This study investigated the applicability to computerized adaptive testing (CAT) data of a differential item functioning (DIF) analysis that involves an empirical Bayes (EB) enhancement of the popular Mantel Haenszel (MH) DIF analysis method. The computerized Law School Admission Test (LSAT) assumed for this study was similar to that currently…
Descriptors: Adaptive Testing, Bayesian Statistics, College Entrance Examinations, Computer Assisted Testing

Zwick, Rebecca; And Others – Journal of Educational Measurement, 1995
In a simulation study of ability and estimation of differential item functioning (DIF) in computerized adaptive tests, Rasch-based DIF statistics were highly correlated with generating DIF, but DIF statistics tended to be slightly smaller than in the three-parameter logistic model analyses. (SLD)
Descriptors: Ability, Adaptive Testing, Computer Assisted Testing, Computer Simulation
Previous Page | Next Page ยป
Pages: 1 | 2