NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1 to 15 of 40 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Arielle Boguslav; Julie Cohen – Journal of Teacher Education, 2024
Teacher preparation programs are increasingly expected to use data on preservice teacher (PST) skills to drive program improvement and provide targeted supports. Observational ratings are especially vital, but also prone to measurement issues. Scores may be influenced by factors unrelated to PSTs' instructional skills, including rater standards.…
Descriptors: Preservice Teachers, Measures (Individuals), Evaluation Problems, Teaching Skills
Peer reviewed Peer reviewed
Direct linkDirect link
Mark White; Matt Ronfeldt – Educational Assessment, 2024
Standardized observation systems seek to reliably measure a specific conceptualization of teaching quality, managing rater error through mechanisms such as certification, calibration, validation, and double-scoring. These mechanisms both support high quality scoring and generate the empirical evidence used to support the scoring inference (i.e.,…
Descriptors: Interrater Reliability, Quality Control, Teacher Effectiveness, Error Patterns
Michelle Herridge – ProQuest LLC, 2021
Evaluation of student written work during summative assessments is an important and critical task for instructors at all educational levels. Nevertheless, few research studies exist that provide insights into how different instructors approach this task. Chemistry faculty (FIs) and graduate student instructors (GSIs) regularly engage in the…
Descriptors: Science Instruction, Chemistry, College Faculty, Teaching Assistants
Peer reviewed Peer reviewed
Direct linkDirect link
Joseph, Gail; Soderberg, Janet S.; Stull, Sara; Cummings, Kevin; McCutchen, Deborah; Han, Rachel J. – Early Education and Development, 2020
Research Findings: This study explores the inter-rater reliability of WaKIDS, Washington State's kindergarten entry assessment (KEA). Specifically, we analyze (1) the extent to which teachers' assessments are in agreement with a master code, (2) how often inaccurate assessment decisions lead to misidentification of school readiness, and (3)…
Descriptors: Interrater Reliability, School Readiness, Kindergarten, Evaluation Problems
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Szafran, Robert F. – Practical Assessment, Research & Evaluation, 2017
Institutional assessment of student learning objectives has become a fact-of-life in American higher education and the Association of American Colleges and Universities' (AAC&U) VALUE Rubrics have become a widely adopted evaluation and scoring tool for student work. As faculty from a variety of disciplines, some less familiar with the…
Descriptors: Interrater Reliability, Case Studies, Scoring Rubrics, Behavioral Objectives
Peer reviewed Peer reviewed
Direct linkDirect link
Parker, Richard I.; Vannest, Kimberly J.; Davis, John L. – Journal of School Psychology, 2013
The use of multi-category scales is increasing for the monitoring of IEP goals, classroom and school rules, and Behavior Improvement Plans (BIPs). Although they require greater inference than traditional data counting, little is known about the inter-rater reliability of these scales. This simulation study examined the performance of nine…
Descriptors: Rating Scales, Scaling, Interrater Reliability, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Csomay, Eniko; Pollard, Elizabeth; Bordelon, Suzanne; Beck, Audrey – Journal of General Education, 2015
Despite the desire of employers to hire those with the critical-thinking and communication skills a general education (GE) program can offer, the value of GE programs is often questioned due to concerns about four-year graduation rates, perceived low immediate economic payoff, and a dearth of evidence to support their efficacy. This article…
Descriptors: General Education, Critical Thinking, Communication Skills, Graduation Rate
Peer reviewed Peer reviewed
Direct linkDirect link
Praetorius, Anna-Katharina; Lenske, Gerlinde; Helmke, Andreas – Learning and Instruction, 2012
Despite considerable interest in the topic of instructional quality in research as well as practice, little is known about the quality of its assessment. Using generalizability analysis as well as content analysis, the present study investigates how reliably and validly instructional quality is measured by observer ratings. Twelve trained raters…
Descriptors: Student Teachers, Interrater Reliability, Content Analysis, Observation
Moffett, David W.; Reid, Barbara K. – Online Submission, 2010
The Investigators studied scoring reliability of Candidates' ten day unit plans of instruction through prescribed action research projects, across three academic years. Scoring of the projects in year one provided opportunities for further refinement of the action research evaluation methods in year two. Across three terms in years one and two…
Descriptors: Research Projects, Action Research, Student Evaluation, Mastery Learning
Peer reviewed Peer reviewed
Direct linkDirect link
Fryer, Marilyn – Creativity Research Journal, 2012
This article explores a number of key issues with regard to the measurement of creativity in the course of conducting psychological research or when applying various evaluation measures. It is argued that, although creativity is a fuzzy concept, it is no more difficult to investigate than other fuzzy concepts people tend to take for granted. At…
Descriptors: Creativity, Educational Research, Psychological Studies, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Farmer, Sybil E.; Wood, Duncan; Swain, Ian D.; Pandyan, Anand D. – International Journal of Rehabilitation Research, 2012
Systematic reviews are used to inform practice, and develop guidelines and protocols. A questionnaire to quantify the risk of bias in systematic reviews, the review paper assessment (RPA) tool, was developed and tested. A search of electronic databases provided a data set of review articles that were then independently reviewed by two assessors…
Descriptors: Outcome Measures, Interrater Reliability, Questionnaires, Literature Reviews
Peer reviewed Peer reviewed
Direct linkDirect link
Clarkeburn, Henriikka; Kettula, Kirsi – Teaching in Higher Education, 2012
This study looks at the fairness of assessing learning journals both as the fairness in creating a valid and robust marking process as well as how different student groups may have unfair disadvantages in performing well in reflective assessment tasks. The fairness of a marking process is discussed through reflecting on the practical process and…
Descriptors: Student Evaluation, Reflection, Summative Evaluation, Formative Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Baker, Beverly A. – Assessing Writing, 2010
In high-stakes writing assessments, rater training in the use of a rating scale does not eliminate variability in grade attribution. This realisation has been accompanied by research that explores possible sources of rater variability, such as rater background or rating scale type. However, there has been little consideration thus far of…
Descriptors: Foreign Countries, Writing Evaluation, Writing Tests, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
McKenzie, Robert G. – Learning Disability Quarterly, 2009
The assessment procedures within Response to Intervention (RTI) models have begun to supplant the use of traditional, discrepancy-based frameworks for identifying students with specific learning disabilities (SLD). Many RTI proponents applaud this shift because of perceived shortcomings in utilizing discrepancy as an indicator of SLD. However,…
Descriptors: Intervention, Learning Disabilities, Error of Measurement, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Zhu, Weimo; Rink, Judy; Placek, Judith H.; Graber, Kim C.; Fox, Connie; Fisette, Jennifer L.; Dyson, Ben; Park, Youngsik; Avery, Marybell; Franck, Marian; Raynes, De – Measurement in Physical Education and Exercise Science, 2011
New testing theories, concepts, and psychometric methods (e.g., item response theory, test equating, and item bank) developed during the past several decades have many advantages over previous theories and methods. In spite of their introduction to the field, they have not been fully accepted by physical educators. Further, the manner in which…
Descriptors: Physical Education, Quality Control, Psychometrics, Item Response Theory
Previous Page | Next Page ยป
Pages: 1  |  2  |  3