NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 286 to 300 of 3,122 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ercümend Ersanli; Ali Kilicarslan – Open Journal for Educational Research, 2024
Intelligence has been extensively explored across various disciplines such as psychology, cognitive science, and neurology. Countless scholars have delved into understanding why certain individuals exhibit higher mental acuity and knowledge. Consequently, numerous studies aim to unveil the essence of intelligence and gauge human cognitive…
Descriptors: Intelligence Tests, Nonverbal Tests, Test Construction, Test Validity
Jiyeo Yun – English Teaching, 2023
Studies on automatic scoring systems in writing assessments have also evaluated the relationship between human and machine scores for the reliability of automated essay scoring systems. This study investigated the magnitudes of indices for inter-rater agreement and discrepancy, especially regarding human and machine scoring, in writing assessment.…
Descriptors: Meta Analysis, Interrater Reliability, Essays, Scoring
Peer reviewed Peer reviewed
Direct linkDirect link
Wind, Stefanie A. – Language Testing, 2019
Differences in rater judgments that are systematically related to construct-irrelevant characteristics threaten the fairness of rater-mediated writing assessments. Accordingly, it is essential that researchers and practitioners examine the degree to which the psychometric quality of rater judgments is comparable across test-taker subgroups.…
Descriptors: Nonparametric Statistics, Interrater Reliability, Differences, Writing Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Nieto, Ricardo; Casabianca, Jodi M. – Journal of Educational Measurement, 2019
Many large-scale assessments are designed to yield two or more scores for an individual by administering multiple sections measuring different but related skills. Multidimensional tests, or more specifically, simple structured tests, such as these rely on multiple multiple-choice and/or constructed responses sections of items to generate multiple…
Descriptors: Tests, Scoring, Responses, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
De Raadt, Alexandra; Warrens, Matthijs J.; Bosker, Roel J.; Kiers, Henk A. L. – Educational and Psychological Measurement, 2019
Cohen's kappa coefficient is commonly used for assessing agreement between classifications of two raters on a nominal scale. Three variants of Cohen's kappa that can handle missing data are presented. Data are considered missing if one or both ratings of a unit are missing. We study how well the variants estimate the kappa value for complete data…
Descriptors: Interrater Reliability, Data, Statistical Analysis, Statistical Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Bais, Frank; Schouten, Barry; Lugtig, Peter; Toepoel, Vera; Arends-Tòth, Judit; Douhou, Salima; Kieruj, Natalia; Morren, Mattijn; Vis, Corrie – Sociological Methods & Research, 2019
Item characteristics can have a significant effect on survey data quality and may be associated with measurement error. Literature on data quality and measurement error is often inconclusive. This could be because item characteristics used for detecting measurement error are not coded unambiguously. In our study, we use a systematic coding…
Descriptors: Foreign Countries, National Surveys, Error of Measurement, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Schack, Edna O.; Dueber, David; Thomas, Jonathan Norris; Fisher, Molly H.; Jong, Cindy – AERA Online Paper Repository, 2019
Scoring of teachers' noticing responses is typically burdened with rater bias and reliance upon interrater consensus. The authors sought to make the scoring process more objective, equitable, and generalizable. The development process began with a description of response characteristics for each professional noticing component disconnected from…
Descriptors: Models, Teacher Evaluation, Observation, Bias
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Elise T. Pas; Lindsay Borden; Katrina J. Debnam; Danielle De Lucia; Catherine P. Bradshaw – Grantee Submission, 2022
Motivational interviewing (MI) is applied in a variety of clinical and coaching models to promote behavior change, with increasing interest in its potential to optimize school-based implementation fidelity. Yet there has been less consideration of fidelity indicators for MI-embedded coaching and links to outcomes. We leveraged secondary data from…
Descriptors: Motivation Techniques, Interviews, Coaching (Performance), Middle School Teachers
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Atilgan, Hakan; Demir, Elif Kübra; Ogretmen, Tuncay; Basokcu, Tahsin Oguz – International Journal of Progressive Education, 2020
It has become a critical question what the reliability level would be when open-ended questions are used in large-scale selection tests. One of the aims of the present study is to determine what the reliability would be in the event that the answers given by test-takers are scored by experts when open-ended short answer questions are used in…
Descriptors: Foreign Countries, Secondary School Students, Test Items, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Shabani, Enayat A.; Panahi, Jaleh – Language Testing in Asia, 2020
The literature on using scoring rubrics in writing assessment denotes the significance of rubrics as practical and useful means to assess the quality of writing tasks. This study tries to investigate the agreement among rubrics endorsed and used for assessing the essay writing tasks by the internationally recognized tests of English language…
Descriptors: Writing Evaluation, Scoring Rubrics, Scores, Interrater Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
McDonald, Jenny; Moskal, Adon Christian Michael; Goodchild, Allen; Stein, Sarah; Terry, Stuart – Assessment & Evaluation in Higher Education, 2020
Student evaluations of teaching and courses (SETs) are part of the fabric of tertiary education and quantitative ratings derived from SETs are highly valued by tertiary institutions. However, many staff do not engage meaningfully with SETs, especially if the process of analysing student feedback is cumbersome or time-consuming. To address this…
Descriptors: Student Evaluation of Teacher Performance, Automation, Content Analysis, Student Reaction
Farley, Jennifer; Duppong Hurley, Kristin; Aitken, A. Angelique – Grantee Submission, 2020
This project explored the reliability and utility of transcription in coding qualitative data across two studies in a program evaluation context. The first study tested the method of direct audio coding, or coding audio files without transcripts, using qualitative data software. The presence and frequency of codes applied in direct audio coding…
Descriptors: Program Implementation, Audio Equipment, Coding, Usability
Randa G. Keeley; Rebecca Alvarado-Alcantar; David W. Keeley – Journal of the American Academy of Special Education Professionals, 2020
This article details the development and statistical validation of the diagnostic, observational tool Assessment of the Inclusion of Students with Special Educational Needs and Disabilities (AISSEND) designed to measure the type, frequency, and duration of inclusive practices implemented within an inclusion classroom. The goal of the research team…
Descriptors: Classroom Observation Techniques, Inclusion, Test Construction, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Álvarez-Díaz, Marcos; Muñiz-Bascón, Luis Magín; Soria-Alemany, Antonio; Veintimilla-Bonet, Alberto; Fernández-Alonso, Rubén – International Journal of Music Education, 2021
Evaluation of music performance in competitive contexts often produces discrepancies between the expert judges. These discrepancies can be reduced by using appropriate rubrics that minimise the differences between judges. The objective of this study was the design and validation of an analytical evaluation rubric, which would allow the most…
Descriptors: Competition, Music Activities, Performance, Scoring Rubrics
Peer reviewed Peer reviewed
Direct linkDirect link
Han, Chao; Zhao, Xiao – Assessment & Evaluation in Higher Education, 2021
The accuracy of peer ratings on students' performance has attracted much attention from higher education researchers. In this study, we attempted to explore the accuracy of peer ratings on the quality of spoken-language interpreting in the context of tertiary-level interpreter training. We sought to understand how different types of peer raters…
Descriptors: Accuracy, Peer Evaluation, Oral Language, Interpretive Skills
Pages: 1  |  ...  |  16  |  17  |  18  |  19  |  20  |  21  |  22  |  23  |  24  |  ...  |  209