NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)10
Since 2006 (last 20 years)34
Audience
Laws, Policies, & Programs
Assessments and Surveys
Center for Epidemiologic…1
What Works Clearinghouse Rating
Showing 1 to 15 of 35 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Johnson, Matthew S.; Sinharay, Sandip – Journal of Educational and Behavioral Statistics, 2020
One common score reported from diagnostic classification assessments is the vector of posterior means of the skill mastery indicators. As with any assessment, it is important to derive and report estimates of the reliability of the reported scores. After reviewing a reliability measure suggested by Templin and Bradshaw, this article suggests three…
Descriptors: Reliability, Probability, Skill Development, Classification
Peer reviewed Peer reviewed
Direct linkDirect link
Ferrando, Pere J.; Lorenzo-Seva, Urbano – Educational and Psychological Measurement, 2019
Measures initially designed to be single-trait often yield data that are compatible with both an essentially unidimensional factor-analysis (FA) solution and a correlated-factors solution. For these cases, this article proposes an approach aimed at providing information for deciding which of the two solutions is the most appropriate and useful.…
Descriptors: Factor Analysis, Computation, Reliability, Goodness of Fit
Peer reviewed Peer reviewed
Direct linkDirect link
Song, Yue; Sun, Feng; Redline, Susan; Wang, Rui – Research Synthesis Methods, 2020
Meta-analyses of clinical trials typically focus on one outcome at a time. However, treatment decision-making depends on an overall assessment of outcomes balancing benefit in various domains and potential risks. This calls for meta-analysis methods for combined outcomes that encompass information from different domains. When individual patient…
Descriptors: Meta Analysis, Patients, Data, Outcomes of Treatment
Bailey, Paul; Emad, Ahmad; Zhang, Ting; Xie, Qingshu; Sikali, Emmanuel – American Institutes for Research, 2018
Correlation analysis has been used widely by researchers and analysts when analyzing large-scale assessment data. Limit research provided reliable methods to estimate various correlations and their standard errors with the complex sampling design and multiple plausible values taken into account. This report introduces the methodology used by the…
Descriptors: Correlation, Educational Assessment, Measurement, Statistical Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Wells, Craig S.; Sireci, Stephen G. – Applied Measurement in Education, 2020
Student growth percentiles (SGPs) are currently used by several states and school districts to provide information about individual students as well as to evaluate teachers, schools, and school districts. For SGPs to be defensible for these purposes, they should be reliable. In this study, we examine the amount of systematic and random error in…
Descriptors: Growth Models, Reliability, Scores, Error Patterns
Peer reviewed Peer reviewed
Direct linkDirect link
Trafimow, David – Educational and Psychological Measurement, 2018
Because error variance alternatively can be considered to be the sum of systematic variance associated with unknown variables and randomness, a tripartite assumption is proposed that total variance in the dependent variable can be partitioned into three variance components. These are variance in the dependent variable that is explained by the…
Descriptors: Statistical Analysis, Correlation, Experiments, Effect Size
Peer reviewed Peer reviewed
Direct linkDirect link
Nicewander, W. Alan – Educational and Psychological Measurement, 2018
Spearman's correction for attenuation (measurement error) corrects a correlation coefficient for measurement errors in either-or-both of two variables, and follows from the assumptions of classical test theory. Spearman's equation removes all measurement error from a correlation coefficient which translates into "increasing the reliability of…
Descriptors: Error of Measurement, Correlation, Sample Size, Computation
Peer reviewed Peer reviewed
Direct linkDirect link
Leckie, George – Journal of Educational and Behavioral Statistics, 2018
The traditional approach to estimating the consistency of school effects across subject areas and the stability of school effects across time is to fit separate value-added multilevel models to each subject or cohort and to correlate the resulting empirical Bayes predictions. We show that this gives biased correlations and these biases cannot be…
Descriptors: Value Added Models, Reliability, Statistical Bias, Computation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
García-Santillán, Arturo; Ochoa-Domínguez, Tomás Elio; Ramos-Hernández, Jésica Josefina – International Electronic Journal of Mathematics Education, 2018
The aim of the study was to measure the level of anxiety towards mathematics among workers in the production area of a factory in the sugar industry. In order to carry out this study, the Muñoz y Mato-Vázquez scale was used (2007) and adapted to the working area. 283 workers from the three different shifts (morning, afternoon and night) were…
Descriptors: Foreign Countries, Mathematics Anxiety, Factor Analysis, Educational Attainment
Peer reviewed Peer reviewed
Direct linkDirect link
Monroe, Scott; Cai, Li – Educational Measurement: Issues and Practice, 2015
Student growth percentiles (SGPs, Betebenner, 2009) are used to locate a student's current score in a conditional distribution based on the student's past scores. Currently, following Betebenner (2009), quantile regression (QR) is most often used operationally to estimate the SGPs. Alternatively, multidimensional item response theory (MIRT) may…
Descriptors: Item Response Theory, Reliability, Growth Models, Computation
Peer reviewed Peer reviewed
Direct linkDirect link
Tesio, Luigi – International Journal of Rehabilitation Research, 2012
Outcome studies in biomedical research usually focus on testing mean changes across samples of subjects and, in so doing, often obscure changes in individuals. These changes, however, may be very informative in studies in which large or homogeneous samples are unavailable and mechanisms of action are still under scrutiny, as is often the case for…
Descriptors: Biomedicine, Correlation, Computation, Behavioral Sciences
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ferrando, Pere J. – Psicologica: International Journal of Methodology and Experimental Psychology, 2015
Test-retest studies for assessing stability and change are widely used in different domains and allow improved or additional individual estimates of interest to be obtained. However, if these estimates are to be validly interpreted the responses given at Time-2 must be free of retest effects, and the fulfilment of this assumption must be…
Descriptors: Item Response Theory, Evaluation Methods, Responses, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Huang, Po-Hsien; Weng, Li-Jen – Multivariate Behavioral Research, 2012
A procedure for estimating the reliability of test scores in the context of ecological momentary assessment (EMA) was proposed to take into account the characteristics of EMA measures. Two commonly used test scores in EMA were considered: the aggregated score (AGGS) and the within-person centered score (WPCS). Conceptually, AGGS and WPCS represent…
Descriptors: Reliability, Scores, Correlation, Computation
Saupe, Joe L.; Eimers, Mardy T. – Association for Institutional Research, 2013
The purpose of this paper is to explore differences in the reliabilities of cumulative college grade point averages (GPAs), estimated for unweighted and weighted, one-semester, 1-year, 2-year, and 4-year GPAs. Using cumulative GPAs for a freshman class at a major university, we estimate internal consistency (coefficient alpha) reliabilities for…
Descriptors: Grade Point Average, College Freshmen, Reliability, Comparative Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Gadermann, Anne M.; Guhn, Martin; Zumbo, Bruno D. – Practical Assessment, Research & Evaluation, 2012
This paper provides a conceptual, empirical, and practical guide for estimating ordinal reliability coefficients for ordinal item response data (also referred to as Likert, Likert-type, ordered categorical, or rating scale item responses). Conventionally, reliability coefficients, such as Cronbach's alpha, are calculated using a Pearson…
Descriptors: Likert Scales, Rating Scales, Reliability, Computation
Previous Page | Next Page »
Pages: 1  |  2  |  3