Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 14 |
Descriptor
Computation | 15 |
Models | 15 |
Item Response Theory | 9 |
Accuracy | 5 |
Maximum Likelihood Statistics | 5 |
Test Items | 5 |
National Competency Tests | 4 |
Reliability | 4 |
Bayesian Statistics | 3 |
Comparative Analysis | 3 |
Equated Scores | 3 |
More ▼ |
Source
ETS Research Report Series | 15 |
Author
Haberman, Shelby J. | 3 |
von Davier, Matthias | 3 |
Ali, Usama S. | 2 |
Antal, Tamás | 2 |
Almond, Russell G. | 1 |
Hartz, Sarah | 1 |
Kandathil, Mathew | 1 |
Li, Deping | 1 |
Moses, Tim | 1 |
Oh, Hyeonjoo J. | 1 |
Oranje, Andreas | 1 |
More ▼ |
Publication Type
Journal Articles | 15 |
Reports - Research | 15 |
Education Level
Elementary Education | 2 |
Grade 8 | 2 |
Junior High Schools | 2 |
Middle Schools | 2 |
Secondary Education | 2 |
Grade 12 | 1 |
Grade 4 | 1 |
High Schools | 1 |
Intermediate Grades | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of… | 4 |
Praxis Series | 1 |
What Works Clearinghouse Rating
von Davier, Matthias – ETS Research Report Series, 2016
This report presents results on a parallel implementation of the expectation-maximization (EM) algorithm for multidimensional latent variable models. The developments presented here are based on code that parallelizes both the E step and the M step of the parallel-E parallel-M algorithm. Examples presented in this report include item response…
Descriptors: Psychometrics, Mathematics, Models, Statistical Analysis
van Rijn, Peter W.; Ali, Usama S. – ETS Research Report Series, 2018
A computer program was developed to estimate speed-accuracy response models for dichotomous items. This report describes how the models are estimated and how to specify data and input files. An example using data from a listening section of an international language test is described to illustrate the modeling approach and features of the computer…
Descriptors: Computer Software, Computation, Reaction Time, Timed Tests
Ali, Usama S.; Walker, Michael E. – ETS Research Report Series, 2014
Two methods are currently in use at Educational Testing Service (ETS) for equating observed item difficulty statistics. The first method involves the linear equating of item statistics in an observed sample to reference statistics on the same items. The second method, or the item response curve (IRC) method, involves the summation of conditional…
Descriptors: Difficulty Level, Test Items, Equated Scores, Causal Models
Wang, Zhen; Yao, Lihua – ETS Research Report Series, 2013
The current study used simulated data to investigate the properties of a newly proposed method (Yao's rater model) for modeling rater severity and its distribution under different conditions. Our study examined the effects of rater severity, distributions of rater severity, the difference between item response theory (IRT) models with rater effect…
Descriptors: Test Format, Test Items, Responses, Computation
Moses, Tim; Oh, Hyeonjoo J. – ETS Research Report Series, 2009
Pseudo Bayes probability estimates are weighted averages of raw and modeled probabilities; these estimates have been studied primarily in nonpsychometric contexts. The purpose of this study was to evaluate pseudo Bayes probability estimates as applied to the estimation of psychometric test score distributions and chained equipercentile equating…
Descriptors: Bayesian Statistics, Computation, Equated Scores, Probability
Oranje, Andreas; Li, Deping; Kandathil, Mathew – ETS Research Report Series, 2009
Several complex sample standard error estimators based on linearization and resampling for the latent regression model of the National Assessment of Educational Progress (NAEP) are studied with respect to design choices such as number of items, number of regressors, and the efficiency of the sample. This paper provides an evaluation of the extent…
Descriptors: Error of Measurement, Computation, Regression (Statistics), National Competency Tests
Haberman, Shelby J. – ETS Research Report Series, 2008
Continuous exponential families are applied to linking forms via a single-group design. In this application, a distribution from the continuous bivariate exponential family is used that has selected moments that match those of the bivariate distribution of scores on the forms to be linked. The selected continuous bivariate distribution then yields…
Descriptors: Equated Scores, Probability, Statistical Distributions, Models
Xu, Xueli; von Davier, Matthias – ETS Research Report Series, 2008
The general diagnostic model (GDM) utilizes located latent classes for modeling a multidimensional proficiency variable. In this paper, the GDM is extended by employing a log-linear model for multiple populations that assumes constraints on parameters across multiple groups. This constrained model is compared to log-linear models that assume…
Descriptors: Comparative Analysis, Models, Computation, National Competency Tests
Antal, Tamás – ETS Research Report Series, 2007
A coordinate-free definition of complex-structure multidimensional item response theory (MIRT) for dichotomously scored items is presented. The point of view taken emphasizes the possibilities and subtleties of understanding MIRT as a multidimensional extension of the classical unidimensional item response theory models. The main theorem of the…
Descriptors: Item Response Theory, Models, Test Items, Computation
Haberman, Shelby J. – ETS Research Report Series, 2007
In item-response theory, if a latent-structure model has an ability variable, then elementary information theory may be employed to provide a criterion for evaluation of the information the test provides concerning ability. This criterion may be considered even in cases in which the latent-structure model is not valid, although interpretation of…
Descriptors: Item Response Theory, Ability, Information Theory, Computation
Antal, Tamás – ETS Research Report Series, 2007
Full account of the latent regression model for the National Assessment of Educational Progress is given. The treatment includes derivation of the EM algorithm, Newton-Raphson method, and the asymptotic standard errors. The paper also features the use of the adaptive Gauss-Hermite numerical integration method as a basic tool to evaluate…
Descriptors: Regression (Statistics), Item Response Theory, National Competency Tests, Evaluation Methods
Haberman, Shelby J. – ETS Research Report Series, 2006
Multinomial-response models are available that correspond implicitly to tests in which a total score is computed as the sum of polytomous item scores. For these models, joint and conditional estimation may be considered in much the same way as for the Rasch model for right-scored tests. As in the Rasch model, joint estimation is only attractive if…
Descriptors: Computation, Models, Test Items, Scores
Almond, Russell G. – ETS Research Report Series, 2007
Over the course of instruction, instructors generally collect a great deal of information about each student. Integrating that information intelligently requires models for how a student's proficiency changes over time. Armed with such models, instructors can "filter" the data--more accurately estimate the student's current proficiency…
Descriptors: Markov Processes, Decision Making, Student Evaluation, Learning Processes
Hartz, Sarah; Roussos, Louis – ETS Research Report Series, 2008
This paper presents the development of the fusion model skills diagnosis system (fusion model system), which can help integrate standardized testing into the learning process with both skills-level examinee parameters for modeling examinee skill mastery and skills-level item parameters, giving information about the diagnostic power of the test.…
Descriptors: Skill Development, Educational Diagnosis, Theory Practice Relationship, Standardized Tests
Sinharay, Sandip; von Davier, Matthias – ETS Research Report Series, 2005
The reporting methods used in large scale assessments such as the National Assessment of Educational Progress (NAEP) rely on a "latent regression model." The first component of the model consists of a "p"-scale IRT measurement model that defines the response probabilities on a set of cognitive items in "p" scales…
Descriptors: National Competency Tests, Regression (Statistics), Predictor Variables, Student Characteristics