Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 1 |
| Since 2017 (last 10 years) | 4 |
| Since 2007 (last 20 years) | 11 |
Descriptor
| Computation | 17 |
| Sampling | 17 |
| National Competency Tests | 10 |
| Probability | 7 |
| Statistical Analysis | 7 |
| Comparative Analysis | 6 |
| Error of Measurement | 6 |
| Grade 8 | 6 |
| Item Response Theory | 5 |
| Grade 4 | 4 |
| Reading Tests | 4 |
| More ▼ | |
Source
| ETS Research Report Series | 17 |
Author
| Qian, Jiahe | 5 |
| Oranje, Andreas | 4 |
| Braun, Henry | 2 |
| Haberman, Shelby J. | 2 |
| Donoghue, John R. | 1 |
| Dorans, Neil J. | 1 |
| Freund, David | 1 |
| Guo, Hongwen | 1 |
| Haberman, Shelby | 1 |
| Hess, Melinda R. | 1 |
| Jenkins, Frank | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 17 |
| Reports - Research | 17 |
Education Level
| Elementary Education | 7 |
| Grade 8 | 6 |
| Junior High Schools | 6 |
| Middle Schools | 6 |
| Secondary Education | 6 |
| Grade 4 | 4 |
| Intermediate Grades | 4 |
| Higher Education | 1 |
| Postsecondary Education | 1 |
Audience
Location
| California | 1 |
| Nevada | 1 |
| New Jersey | 1 |
| United States | 1 |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
| National Assessment of… | 10 |
What Works Clearinghouse Rating
Donoghue, John R.; McClellan, Catherine A.; Hess, Melinda R. – ETS Research Report Series, 2022
When constructed-response items are administered for a second time, it is necessary to evaluate whether the current Time B administration's raters have drifted from the scoring of the original administration at Time A. To study this, Time A papers are sampled and rescored by Time B scorers. Commonly the scores are compared using the proportion of…
Descriptors: Item Response Theory, Test Construction, Scoring, Testing
Qian, Jiahe – ETS Research Report Series, 2020
The finite population correction (FPC) factor is often used to adjust variance estimators for survey data sampled from a finite population without replacement. As a replicated resampling approach, the jackknife approach is usually implemented without the FPC factor incorporated in its variance estimates. A paradigm is proposed to compare the…
Descriptors: Computation, Sampling, Data, Statistical Analysis
Jewsbury, Paul A. – ETS Research Report Series, 2019
When an assessment undergoes changes to the administration or instrument, bridge studies are typically used to try to ensure comparability of scores before and after the change. Among the most common and powerful is the common population linking design, with the use of a linear transformation to link scores to the metric of the original…
Descriptors: Evaluation Research, Scores, Error Patterns, Error of Measurement
Qian, Jiahe – ETS Research Report Series, 2017
The variance formula derived for a two-stage sampling design without replacement employs the joint inclusion probabilities in the first-stage selection of clusters. One of the difficulties encountered in data analysis is the lack of information about such joint inclusion probabilities. One way to solve this issue is by applying Hájek's…
Descriptors: Mathematical Formulas, Computation, Sampling, Research Design
Lu, Ru; Haberman, Shelby; Guo, Hongwen; Liu, Jinghua – ETS Research Report Series, 2015
In this study, we apply jackknifing to anchor items to evaluate the impact of anchor selection on equating stability. In an ideal world, the choice of anchor items should have little impact on equating results. When this ideal does not correspond to reality, selection of anchor items can strongly influence equating results. This influence does not…
Descriptors: Test Construction, Equated Scores, Test Items, Sampling
Qian, Jiahe; Jiang, Yanming; von Davier, Alina A. – ETS Research Report Series, 2013
Several factors could cause variability in item response theory (IRT) linking and equating procedures, such as the variability across examinee samples and/or test items, seasonality, regional differences, native language diversity, gender, and other demographic variables. Hence, the following question arises: Is it possible to select optimal…
Descriptors: Item Response Theory, Test Items, Sampling, True Scores
Oranje, Andreas; Li, Deping; Kandathil, Mathew – ETS Research Report Series, 2009
Several complex sample standard error estimators based on linearization and resampling for the latent regression model of the National Assessment of Educational Progress (NAEP) are studied with respect to design choices such as number of items, number of regressors, and the efficiency of the sample. This paper provides an evaluation of the extent…
Descriptors: Error of Measurement, Computation, Regression (Statistics), National Competency Tests
Haberman, Shelby J. – ETS Research Report Series, 2008
Continuous exponential families may be employed to find continuous distributions with the same initial moments as the discrete distributions encountered in typical applications of classical equating. These continuous distributions provide distribution functions and quantile functions that may be employed in equating. To illustrate, an application…
Descriptors: Equated Scores, Statistical Distributions, Probability, Computation
Braun, Henry; Zhang, Jinming; Vezzu, Sailesh – ETS Research Report Series, 2008
At present, although the percentages of students with disabilities (SDs) and/or students who are English language learners (ELL) excluded from a NAEP administration are reported, no statistical adjustment is made for these excluded students in the calculation of NAEP results. However, the exclusion rates for both SD and ELL students vary…
Descriptors: Research Methodology, Computation, Disabilities, English Language Learners
Haberman, Shelby J. – ETS Research Report Series, 2004
Criteria for prediction of multinomial responses are examined in terms of estimation bias. Logarithmic penalty and least squares are quite similar in behavior but quite different from maximum probability. The differences ultimately reflect deficiencies in the behavior of the criterion of maximum probability.
Descriptors: Probability, Prediction, Classification, Computation
Braun, Henry; Qian, Jiahe – ETS Research Report Series, 2008
This report describes the derivation and evaluation of a method for comparing the performance standards for public school students set by different states. It is based on an approach proposed by McLaughlin and associates, which constituted an innovative attempt to resolve the confusion and concern that occurs when very different proportions of…
Descriptors: State Standards, Comparative Analysis, Public Schools, National Competency Tests
Oranje, Andreas – ETS Research Report Series, 2006
Confidence intervals are an important tool to indicate uncertainty of estimates and to give an idea of probable values of an estimate if a different sample from the population was drawn or a different sample of measures was used. Standard symmetric confidence intervals for proportion estimates based on a normal approximation can yield bounds…
Descriptors: Computation, Statistical Analysis, National Competency Tests, Comparative Analysis
Livingston, Samuel A.; Dorans, Neil J. – ETS Research Report Series, 2004
This paper describes an approach to item analysis that is based on the estimation of a set of response curves for each item. The response curves show, at a glance, the difficulty and the discriminating power of the item and the popularity of each distractor, at any level of the criterion variable (e.g., total score). The curves are estimated by…
Descriptors: Item Analysis, Computation, Difficulty Level, Test Items
Oranje, Andreas – ETS Research Report Series, 2006
A multitude of methods has been proposed to estimate the sampling variance of ratio estimates in complex samples (Wolter, 1985). Hansen and Tepping (1985) studied some of those variance estimators and found that a high coefficient of variation (CV) of the denominator of a ratio estimate is indicative of a biased estimate of the standard error of a…
Descriptors: Statistical Analysis, Computation, Sampling, Statistical Bias
Oranje, Andreas; Freund, David; Lin, Mei-jang; Tang, Yuxin – ETS Research Report Series, 2007
In this paper, a data perturbation method for minimizing the possibility of disclosure of participants' identities on a survey is described in the context of the National Assessment of Educational Progress (NAEP). The method distinguishes itself from most approaches because of the presence of cognitive tasks. Hence, a data edit should have minimal…
Descriptors: Student Surveys, Risk, National Competency Tests, Data Analysis
Previous Page | Next Page »
Pages: 1 | 2
Peer reviewed
