Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 0 |
| Since 2007 (last 20 years) | 8 |
Descriptor
| Computation | 9 |
| Statistical Analysis | 9 |
| Equated Scores | 6 |
| Accuracy | 4 |
| Error of Measurement | 4 |
| Simulation | 4 |
| Comparative Analysis | 3 |
| Sample Size | 3 |
| Regression (Statistics) | 2 |
| Statistical Significance | 2 |
| Test Bias | 2 |
| More ▼ | |
Source
| ETS Research Report Series | 3 |
| Journal of Educational and… | 3 |
| Journal of Educational… | 2 |
| Educational Testing Service | 1 |
Author
| Moses, Tim | 9 |
| Holland, Paul | 2 |
| Miao, Jing | 2 |
| Dorans, Neil | 1 |
| Dorans, Neil J. | 1 |
| Holland, Paul W. | 1 |
| Kim, Sooyeon | 1 |
| Yoo, Hanwook | 1 |
| Zhang, Wenmin | 1 |
Publication Type
| Journal Articles | 8 |
| Reports - Evaluative | 4 |
| Reports - Research | 4 |
| Reports - Descriptive | 1 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Kim, Sooyeon; Moses, Tim; Yoo, Hanwook – Journal of Educational Measurement, 2015
This inquiry is an investigation of item response theory (IRT) proficiency estimators' accuracy under multistage testing (MST). We chose a two-stage MST design that includes four modules (one at Stage 1, three at Stage 2) and three difficulty paths (low, middle, high). We assembled various two-stage MST panels (i.e., forms) by manipulating two…
Descriptors: Comparative Analysis, Item Response Theory, Computation, Accuracy
Moses, Tim; Zhang, Wenmin – Journal of Educational and Behavioral Statistics, 2011
The purpose of this article was to extend the use of standard errors for equated score differences (SEEDs) to traditional equating functions. The SEEDs are described in terms of their original proposal for kernel equating functions and extended so that SEEDs for traditional linear and traditional equipercentile equating functions can be computed.…
Descriptors: Equated Scores, Error Patterns, Evaluation Research, Statistical Analysis
Moses, Tim; Miao, Jing; Dorans, Neil J. – Journal of Educational and Behavioral Statistics, 2010
In this study, the accuracies of four strategies were compared for estimating conditional differential item functioning (DIF), including raw data, logistic regression, log-linear models, and kernel smoothing. Real data simulations were used to evaluate the estimation strategies across six items, DIF and No DIF situations, and four sample size…
Descriptors: Test Bias, Statistical Analysis, Computation, Comparative Analysis
Moses, Tim; Holland, Paul W. – Journal of Educational Measurement, 2009
In this study, we compared 12 statistical strategies proposed for selecting loglinear models for smoothing univariate test score distributions and for enhancing the stability of equipercentile equating functions. The major focus was on evaluating the effects of the selection strategies on equating function accuracy. Selection strategies' influence…
Descriptors: Equated Scores, Selection, Statistical Analysis, Models
Moses, Tim; Miao, Jing; Dorans, Neil – Educational Testing Service, 2010
This study compared the accuracies of four differential item functioning (DIF) estimation methods, where each method makes use of only one of the following: raw data, logistic regression, loglinear models, or kernel smoothing. The major focus was on the estimation strategies' potential for estimating score-level, conditional DIF. A secondary focus…
Descriptors: Test Bias, Statistical Analysis, Computation, Scores
Moses, Tim; Holland, Paul – ETS Research Report Series, 2008
The purpose of this paper is to extend von Davier, Holland, and Thayer's (2004b) framework of kernel equating so that it can incorporate raw data and traditional equipercentile equating methods. One result of this more general framework is that previous equating methodology research can be viewed more comprehensively. Another result is that the…
Descriptors: Equated Scores, Error of Measurement, Statistical Analysis, Computation
Moses, Tim – Journal of Educational and Behavioral Statistics, 2008
Equating functions are supposed to be population invariant, meaning that the choice of subpopulation used to compute the equating function should not matter. The extent to which equating functions are population invariant is typically assessed in terms of practical difference criteria that do not account for equating functions' sampling…
Descriptors: Equated Scores, Error of Measurement, Sampling, Evaluation Methods
Moses, Tim; Holland, Paul – ETS Research Report Series, 2007
The purpose of this study was to empirically evaluate the impact of loglinear presmoothing accuracy on equating bias and variability across chained and post-stratification equating methods, kernel and percentile-rank continuization methods, and sample sizes. The results of evaluating presmoothing on equating accuracy generally agreed with those of…
Descriptors: Equated Scores, Statistical Analysis, Accuracy, Sample Size
Moses, Tim – ETS Research Report Series, 2006
Population invariance is an important requirement of test equating. An equating function is said to be population invariant when the choice of (sub)population used to compute the equating function does not matter. In recent studies, the extent to which equating functions are population invariant is typically addressed in terms of practical…
Descriptors: Equated Scores, Computation, Error of Measurement, Statistical Analysis

Peer reviewed
Direct link
