Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 7 |
Descriptor
Methods | 7 |
Item Response Theory | 4 |
Computation | 3 |
Models | 3 |
Scores | 3 |
Comparative Analysis | 2 |
Data | 2 |
Licensing Examinations… | 2 |
Prediction | 2 |
Reliability | 2 |
Simulation | 2 |
More ▼ |
Source
Educational Testing Service | 2 |
Applied Measurement in… | 1 |
Applied Psychological… | 1 |
Educational Measurement:… | 1 |
Educational and Psychological… | 1 |
Journal of Educational and… | 1 |
Author
Sinharay, Sandip | 7 |
Haberman, Shelby J. | 2 |
Puhan, Gautam | 2 |
Haberman, Shelby | 1 |
Holland, Paul W. | 1 |
Larkin, Kevin | 1 |
Levy, Roy | 1 |
Mislevy, Robert J. | 1 |
Wainer, Howard | 1 |
von Davier, Matthias | 1 |
Publication Type
Journal Articles | 5 |
Reports - Evaluative | 3 |
Reports - Descriptive | 2 |
Reports - Research | 2 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Sinharay, Sandip – Journal of Educational and Behavioral Statistics, 2015
Person-fit assessment may help the researcher to obtain additional information regarding the answering behavior of persons. Although several researchers examined person fit, there is a lack of research on person-fit assessment for mixed-format tests. In this article, the lz statistic and the ?2 statistic, both of which have been used for tests…
Descriptors: Test Format, Goodness of Fit, Item Response Theory, Bayesian Statistics
Sinharay, Sandip; Haberman, Shelby J.; Wainer, Howard – Educational and Psychological Measurement, 2011
There are several techniques that increase the precision of subscores by borrowing information from other parts of the test. These techniques have been criticized on validity grounds in several of the recent publications. In this note, the authors question the argument used in these publications and suggest both inherent limits to the validity…
Descriptors: Scores, Methods, Validity, Reliability
Sinharay, Sandip; Puhan, Gautam; Haberman, Shelby J. – Educational Measurement: Issues and Practice, 2011
The purpose of this ITEMS module is to provide an introduction to subscores. First, examples of subscores from an operational test are provided. Then, a review of methods that can be used to examine if subscores have adequate psychometric quality is provided. It is demonstrated, using results from operational and simulated data, that subscores…
Descriptors: Scores, Psychometrics, Tests, Data
Puhan, Gautam; Sinharay, Sandip; Haberman, Shelby; Larkin, Kevin – Applied Measurement in Education, 2010
Will subscores provide additional information than what is provided by the total score? Is there a method that can estimate more trustworthy subscores than observed subscores? To answer the first question, this study evaluated whether the true subscore was more accurately predicted by the observed subscore or total score. To answer the second…
Descriptors: Licensing Examinations (Professions), Scores, Computation, Methods
Levy, Roy; Mislevy, Robert J.; Sinharay, Sandip – Applied Psychological Measurement, 2009
If data exhibit multidimensionality, key conditional independence assumptions of unidimensional models do not hold. The current work pursues posterior predictive model checking, a flexible family of model-checking procedures, as a tool for criticizing models due to unaccounted for dimensions in the context of item response theory. Factors…
Descriptors: Item Response Theory, Models, Methods, Simulation
von Davier, Matthias; Sinharay, Sandip – Educational Testing Service, 2009
This paper presents an application of a stochastic approximation EM-algorithm using a Metropolis-Hastings sampler to estimate the parameters of an item response latent regression model. Latent regression models are extensions of item response theory (IRT) to a 2-level latent variable model in which covariates serve as predictors of the…
Descriptors: Item Response Theory, Regression (Statistics), Models, Methods
Sinharay, Sandip; Holland, Paul W. – Educational Testing Service, 2008
The nonequivalent groups with anchor test (NEAT) design involves missing data that are missing by design. Three popular equating methods that can be used with a NEAT design are the poststratification equating method, the chain equipercentile equating method, and the item-response-theory observed-score-equating method. These three methods each…
Descriptors: Equated Scores, Test Items, Item Response Theory, Data