NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Type
Reports - Descriptive27
Journal Articles24
Opinion Papers1
Audience
Researchers1
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 27 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Gyamfi, Abraham; Acquaye, Rosemary – Acta Educationis Generalis, 2023
Introduction: Item response theory (IRT) has received much attention in validation of assessment instrument because it allows the estimation of students' ability from any set of the items. Item response theory allows the difficulty and discrimination levels of each item on the test to be estimated. In the framework of IRT, item characteristics are…
Descriptors: Item Response Theory, Models, Test Items, Difficulty Level
Peer reviewed Peer reviewed
Direct linkDirect link
Zhao, Xin; Coxe, Stefany; Sibley, Margaret H.; Zulauf-McCurdy, Courtney; Pettit, Jeremy W. – Prevention Science, 2023
There has been increasing interest in applying integrative data analysis (IDA) to analyze data across multiple studies to increase sample size and statistical power. Measures of a construct are frequently not consistent across studies. This article provides a tutorial on the complex decisions that occur when conducting harmonization of measures…
Descriptors: Data Analysis, Sample Size, Decision Making, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Marcoulides, George A.; Huber, Chuck – Measurement: Interdisciplinary Research and Perspectives, 2020
It is demonstrated that the popular three-parameter logistic model can lead to markedly inaccurate individual ability level estimates for mixture populations. A theoretically and empirically important setting is initially considered where (a) in one of two subpopulations (latent classes) the two-parameter logistic model holds for each item in a…
Descriptors: Item Response Theory, Models, Measurement Techniques, Item Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Finch, Holmes – Practical Assessment, Research & Evaluation, 2022
Researchers in many disciplines work with ranking data. This data type is unique in that it is often deterministic in nature (the ranks of items "k"-1 determine the rank of item "k"), and the difference in a pair of rank scores separated by "k" units is equivalent regardless of the actual values of the two ranks in…
Descriptors: Data Analysis, Statistical Inference, Models, College Faculty
Gin, Brian; Sim, Nicholas; Skrondal, Anders; Rabe-Hesketh, Sophia – Grantee Submission, 2020
We propose a dyadic Item Response Theory (dIRT) model for measuring interactions of pairs of individuals when the responses to items represent the actions (or behaviors, perceptions, etc.) of each individual (actor) made within the context of a dyad formed with another individual (partner). Examples of its use include the assessment of…
Descriptors: Item Response Theory, Generalization, Item Analysis, Problem Solving
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Aybek, Eren Can; Demirtasli, R. Nukhet – International Journal of Research in Education and Science, 2017
This article aims to provide a theoretical framework for computerized adaptive tests (CAT) and item response theory models for polytomous items. Besides that, it aims to introduce the simulation and live CAT software to the related researchers. Computerized adaptive test algorithm, assumptions of item response theory models, nominal response…
Descriptors: Computer Assisted Testing, Adaptive Testing, Item Response Theory, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Maydeu-Olivares, Alberto – Measurement: Interdisciplinary Research and Perspectives, 2013
In this rejoinder, Maydeu-Olivares states that, in item response theory (IRT) measurement applications, the application of goodness-of-fit (GOF) methods informs researchers of the discrepancy between the model and the data being fitted (the room for improvement). By routinely reporting the GOF of IRT models, together with the substantive results…
Descriptors: Goodness of Fit, Models, Evaluation Methods, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Long, Caroline; Dunne, Tim; Mokoena, Gabriel – Perspectives in Education, 2014
The rationale for the introduction of standards in the United States in the late 1980s was that the quality of education would improve. Assessment instruments in the form of written tests were constructed in order to perform a monitoring function. The introduction of standards and the associated monitoring have been replicated in South Africa. It…
Descriptors: Models, Evaluation Methods, Classroom Environment, Standards
Goldhaber, Dan; Chaplin, Duncan – Center for Education Data & Research, 2012
In a provocative and influential paper, Jesse Rothstein (2010) finds that standard value added models (VAMs) suggest implausible future teacher effects on past student achievement, a finding that obviously cannot be viewed as causal. This is the basis of a falsification test (the Rothstein falsification test) that appears to indicate bias in VAM…
Descriptors: School Effectiveness, Teacher Effectiveness, Achievement Gains, Statistical Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Kreiner, Svend – Applied Psychological Measurement, 2011
To rule out the need for a two-parameter item response theory (IRT) model during item analysis by Rasch models, it is important to check the Rasch model's assumption that all items have the same item discrimination. Biserial and polyserial correlation coefficients measuring the association between items and restscores are often used in an informal…
Descriptors: Item Analysis, Correlation, Item Response Theory, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Ketterlin-Geller, Leanne R.; Yovanoff, Paul; Jung, EunJu; Liu, Kimy; Geller, Josh – Educational Assessment, 2013
In this article, we highlight the need for a precisely defined construct in score-based validation and discuss the contribution of cognitive theories to accurately and comprehensively defining the construct. We propose a framework for integrating cognitively based theoretical and empirical evidence to specify and evaluate the construct. We apply…
Descriptors: Test Validity, Construct Validity, Scores, Evidence
Peer reviewed Peer reviewed
Direct linkDirect link
Wilson, Mark; Moore, Stephen – Language Testing, 2011
This paper provides a summary of a novel and integrated way to think about the item response models (most often used in measurement applications in social science areas such as psychology, education, and especially testing of various kinds) from the viewpoint of the statistical theory of generalized linear and nonlinear mixed models. In addition,…
Descriptors: Reading Comprehension, Testing, Social Sciences, Item Response Theory
Harris, Douglas N. – Phi Delta Kappan, 2010
Current value-added models for teacher accountability are better than models based only on student achievement, but they have their weakness. They are subject to systematic and random error, as are all measures, and there are concerns about the tests used for the measurements. However, value-added models are better than the alternatives at the…
Descriptors: School Effectiveness, Error of Measurement, Achievement Gains, Academic Achievement
Peer reviewed Peer reviewed
Direct linkDirect link
Ferrando, Pere J.; Lorenzo-Seva, Urbano; Chico, Eliseo – Structural Equation Modeling: A Multidisciplinary Journal, 2009
This article proposes procedures for simultaneously assessing and controlling acquiescence and social desirability in questionnaire items. The procedures are based on a semi-restricted factor-analytic tridimensional model, and can be used with binary, graded-response, or more continuous items. We discuss procedures for fitting the model (item…
Descriptors: Factor Analysis, Response Style (Tests), Questionnaires, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Bartholomew, David J.; Deary, Ian J.; Lawn, Martin – Psychological Review, 2009
Modern factor analysis is the outgrowth of Spearman's original "2-factor" model of intelligence, according to which a mental test score is regarded as the sum of a general factor and a specific factor. As early as 1914, Godfrey Thomson realized that the data did not require this interpretation and he demonstrated this by proposing what became…
Descriptors: Psychological Testing, Factor Analysis, Intelligence Tests, Cognitive Ability
Previous Page | Next Page ยป
Pages: 1  |  2