NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 6 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Bardhoshi, Gerta; Erford, Bradley T. – Measurement and Evaluation in Counseling and Development, 2017
Precision is a key facet of test development, with score reliability determined primarily according to the types of error one wants to approximate and demonstrate. This article identifies and discusses several primary forms of reliability estimation: internal consistency (i.e., split-half, KR-20, a), test-retest, alternate forms, interscorer, and…
Descriptors: Scores, Test Reliability, Accuracy, Pretests Posttests
Peer reviewed Peer reviewed
Direct linkDirect link
Gaches, Sonya; Hill, Diana – Journal of Curriculum and Pedagogy, 2017
The CLASS (classroom assessment scoring system) has become integrally linked with quality rating and improvement systems (QRIS) throughout the United States and other international locations. This relationship reinforces the neoliberal consumer-based perspectives of quality and devalues localized perspectives. This article challenges the notion of…
Descriptors: Classroom Techniques, Evaluation Methods, Evaluation Research, Evaluation Problems
Peer reviewed Peer reviewed
Direct linkDirect link
Van Hecke, Tanja – Teaching Mathematics and Its Applications, 2015
Optimal assessment tools should measure in a limited time the knowledge of students in a correct and unbiased way. A method for automating the scoring is multiple choice scoring. This article compares scoring methods from a probabilistic point of view by modelling the probability to pass: the number right scoring, the initial correction (IC) and…
Descriptors: Multiple Choice Tests, Error Correction, Grading, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Attali, Yigal – Applied Psychological Measurement, 2011
Recently, Attali and Powers investigated the usefulness of providing immediate feedback on the correctness of answers to constructed response questions and the opportunity to revise incorrect answers. This article introduces an item response theory (IRT) model for scoring revised responses to questions when several attempts are allowed. The model…
Descriptors: Feedback (Response), Item Response Theory, Models, Error Correction
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ramineni, Chaitanya; Trapani, Catherine S.; Williamson, David M.; Davey, Tim; Bridgeman, Brent – ETS Research Report Series, 2012
Automated scoring models for the "e-rater"® scoring engine were built and evaluated for the "GRE"® argument and issue-writing tasks. Prompt-specific, generic, and generic with prompt-specific intercept scoring models were built and evaluation statistics such as weighted kappas, Pearson correlations, standardized difference in…
Descriptors: Scoring, Test Scoring Machines, Automation, Models
Laitsch, Dan – Association for Supervision and Curriculum Development, 2005
Standardized testing plays an increasingly important role in the lives of today's students and educators. The U.S. No Child Left Behind Act (NCLB) requires assessment in math and literacy in grades 3-8 and 10 and, as of 2007-08, in science once in grades 3-5, 6-9, and 10-12. Based on National Center for Education Statistics enrollment projections,…
Descriptors: Testing, Standardized Tests, Enrollment Projections, Accountability