NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20260
Since 20250
Since 2022 (last 5 years)0
Since 2017 (last 10 years)3
Since 2007 (last 20 years)13
Audience
Laws, Policies, & Programs
Assessments and Surveys
Graduate Record Examinations1
What Works Clearinghouse Rating
Showing 1 to 15 of 19 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Wyse, Adam E. – Practical Assessment, Research & Evaluation, 2018
One common modification to the Angoff standard-setting method is to have panelists round their ratings to the nearest 0.05 or 0.10 instead of 0.01. Several reasons have been offered as to why it may make sense to have panelists round their ratings to the nearest 0.05 or 0.10. In this article, we examine one reason that has been suggested, which is…
Descriptors: Interrater Reliability, Evaluation Criteria, Scoring Formulas, Achievement Rating
Peer reviewed Peer reviewed
Direct linkDirect link
Raczynski, Kevin; Cohen, Allan – Applied Measurement in Education, 2018
The literature on Automated Essay Scoring (AES) systems has provided useful validation frameworks for any assessment that includes AES scoring. Furthermore, evidence for the scoring fidelity of AES systems is accumulating. Yet questions remain when appraising the scoring performance of AES systems. These questions include: (a) which essays are…
Descriptors: Essay Tests, Test Scoring Machines, Test Validity, Evaluators
Peer reviewed Peer reviewed
Direct linkDirect link
Bardhoshi, Gerta; Erford, Bradley T. – Measurement and Evaluation in Counseling and Development, 2017
Precision is a key facet of test development, with score reliability determined primarily according to the types of error one wants to approximate and demonstrate. This article identifies and discusses several primary forms of reliability estimation: internal consistency (i.e., split-half, KR-20, a), test-retest, alternate forms, interscorer, and…
Descriptors: Scores, Test Reliability, Accuracy, Pretests Posttests
Guskey, Thomas R.; Jung, Lee Ann – Educational Leadership, 2016
Many educators consider grades calculated from statistical algorithms more accurate, objective, and reliable than grades they calculate themselves. But in this research, the authors first asked teachers to use their professional judgment to choose a summary grade for hypothetical students. When the researchers compared the teachers' grade with the…
Descriptors: Grading, Computer Assisted Testing, Interrater Reliability, Grades (Scholastic)
Peer reviewed Peer reviewed
Direct linkDirect link
Tarricone, Pina; Newhouse, C. Paul – Australian Educational Researcher, 2016
Traditional moderation of student assessments is often carried out with groups of teachers working face-to-face in a specified location making judgements concerning the quality of representations of achievement. This traditional model has relied little on modern information communications technologies and has been logistically challenging. We…
Descriptors: Visual Arts, Art Education, Art Materials, Alternative Assessment
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Temel, Gülhan Orekici; Erdogan, Semra; Selvi, Hüseyin; Kaya, Irem Ersöz – Educational Sciences: Theory and Practice, 2016
Studies based on longitudinal data focus on the change and development of the situation being investigated and allow for examining cases regarding education, individual development, cultural change, and socioeconomic improvement in time. However, as these studies require taking repeated measures in different time periods, they may include various…
Descriptors: Investigations, Sample Size, Longitudinal Studies, Interrater Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Herbert, Ian P.; Joyce, John; Hassall, Trevor – Accounting Education, 2014
The design, delivery and assessment of a complete educational scheme, such as a degree programme or a professional qualification course, is a complex matter. Maintaining alignment between the stated aims of the curriculum and the scoring of student achievement is an overarching concern. The potential for drift across individual aspects of an…
Descriptors: Higher Education, Student Evaluation, Communities of Practice, Interrater Reliability
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Beltrán, Jorge – Working Papers in TESOL & Applied Linguistics, 2016
In the assessment of aural skills of second language learners, the study of the inclusion of visual stimuli has almost exclusively been conducted in the context of listening assessment. While the inclusion of contextual information in test input has been advocated for by numerous researchers (Ockey, 2010), little has been said regarding the…
Descriptors: Achievement Tests, Speech Skills, Speech Tests, Second Language Learning
Hixson, Nate; Rhudy, Vaughn – West Virginia Department of Education, 2013
Student responses to the West Virginia Educational Standards Test (WESTEST) 2 Online Writing Assessment are scored by a computer-scoring engine. The scoring method is not widely understood among educators, and there exists a misperception that it is not comparable to hand scoring. To address these issues, the West Virginia Department of Education…
Descriptors: Scoring Formulas, Scoring Rubrics, Interrater Reliability, Test Scoring Machines
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Samad, Arshad Abd; bt Ahmad, Zamzam – Advances in Language and Literary Studies, 2012
Raimes (1983) has identified nine components necessary to produce a piece of writing that is clear, fluent and effective. These are also the aspects that are considered when assessing writing. The common practice is to have raters score the essays and they are provided with a rating scale for this purpose. A training and practice session is also…
Descriptors: Writing Evaluation, Writing Achievement, Interrater Reliability, Scoring Formulas
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ramineni, Chaitanya; Trapani, Catherine S.; Williamson, David M.; Davey, Tim; Bridgeman, Brent – ETS Research Report Series, 2012
Automated scoring models for the "e-rater"® scoring engine were built and evaluated for the "GRE"® argument and issue-writing tasks. Prompt-specific, generic, and generic with prompt-specific intercept scoring models were built and evaluation statistics such as weighted kappas, Pearson correlations, standardized difference in…
Descriptors: Scoring, Test Scoring Machines, Automation, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Barkaoui, Khaled – Assessment in Education: Principles, Policy & Practice, 2011
This study examined the effects of marking method and rater experience on ESL (English as a Second Language) essay test scores and rater performance. Each of 31 novice and 29 experienced raters rated a sample of ESL essays both holistically and analytically. Essay scores were analysed using a multi-faceted Rasch model to compare test-takers'…
Descriptors: Writing Evaluation, Writing Tests, Essay Tests, Interrater Reliability
Peer reviewed Peer reviewed
Hughes, Garry L.; Prien, Erich P. – Personnel Psychology, 1986
Investigated psychometric properties of three methods of scoring a Mixed Standard Scale performance evaluation: a patterned procedure, simple nonpatterned scoring procedure and procedure assigning differential weights to statements on the basis of scale values provided by subject matter experts. Found no differences in the score distribution…
Descriptors: Evaluation Methods, Interrater Reliability, Scoring, Scoring Formulas
Berger, Peter N. – Teaching and Learning Literature with Children and Young Adults, 1997
Discusses problems with scoring reliability of the Vermont Education Department's writing portfolio test, particularly the difficulties teachers face in agreeing upon scoring criteria. (PA)
Descriptors: Elementary Secondary Education, Interrater Reliability, Portfolio Assessment, Portfolios (Background Materials)
Peer reviewed Peer reviewed
Upshur, John A.; Turner, Carolyn E. – ELT Journal, 1995
Reviews the place of rating scales in second-language measurement and summarizes some of the problems associated with them. Standard and alternative scales were studied. High agreement among raters can be achieved even under conditions not favorable to high interrater reliability. The full range of score categories are effectively utilized. (17…
Descriptors: Evaluation Problems, Interrater Reliability, Language Tests, Measurement Techniques
Previous Page | Next Page »
Pages: 1  |  2