Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 2 |
| Since 2007 (last 20 years) | 14 |
Descriptor
| Comparative Analysis | 24 |
| Interrater Reliability | 24 |
| Interviews | 24 |
| Language Tests | 7 |
| Foreign Countries | 6 |
| Evaluation Methods | 5 |
| Oral Language | 5 |
| Second Language Instruction | 5 |
| Test Reliability | 5 |
| Language Proficiency | 4 |
| Rating Scales | 4 |
| More ▼ | |
Source
Author
Publication Type
| Journal Articles | 16 |
| Reports - Research | 16 |
| Reports - Evaluative | 5 |
| Dissertations/Theses -… | 3 |
| Speeches/Meeting Papers | 3 |
| Collected Works - Serials | 1 |
| Tests/Questionnaires | 1 |
Education Level
| Higher Education | 7 |
| Postsecondary Education | 5 |
| Elementary Education | 3 |
| Secondary Education | 2 |
| Elementary Secondary Education | 1 |
| Grade 3 | 1 |
| Grade 4 | 1 |
| Grade 5 | 1 |
| Grade 6 | 1 |
| Middle Schools | 1 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Mandy, William; Clarke, Kiri; McKenner, Michele; Strydom, Andre; Crabtree, Jason; Lai, Meng-Chuan; Allison, Carrie; Baron-Cohen, Simon; Skuse, David – Journal of Autism and Developmental Disorders, 2018
We developed a brief, informant-report interview for assessing autism spectrum conditions (ASC) in adults, called the Developmental, Dimensional and Diagnostic Interview-Adult Version (3Di-Adult); and completed a preliminary evaluation. Informant reports were collected for participants with ASC (n = 39), a non-clinical comparison group (n = 29)…
Descriptors: Autism, Pervasive Developmental Disorders, Adults, Diagnostic Tests
Dogan, C. Deha; Uluman, Müge – Educational Sciences: Theory and Practice, 2017
The aim of this study was to determine the extent at which graded-category rating scales and rubrics contribute to inter-rater reliability. The research was designed as a correlational study. Study group consisted of 82 students attending sixth grade and three writing course teachers in a private elementary school. A performance task was…
Descriptors: Comparative Analysis, Scoring Rubrics, Rating Scales, Interrater Reliability
Shubert, Christopher W.; Meredith, Dawn C. – Physical Review Special Topics - Physics Education Research, 2015
Students' epistemologies affect how and what they learn: do they believe physics is a list of equations, or a coherent and sensible description of the physical world? In order to study these epistemologies as part of curricular assessment, we adopt the resources framework, which posits that students have many productive epistemological resources…
Descriptors: Epistemology, Recall (Psychology), Physics, Educational Environment
Ahmadi, Alireza; Sadeghi, Elham – Language Assessment Quarterly, 2016
In the present study we investigated the effect of test format on oral performance in terms of test scores and discourse features (accuracy, fluency, and complexity). Moreover, we explored how the scores obtained on different test formats relate to such features. To this end, 23 Iranian EFL learners participated in three test formats of monologue,…
Descriptors: Oral Language, Comparative Analysis, Language Fluency, Accuracy
Schleigh, Sharon Price; Clark, Douglas B.; Menekse, Muhsin – International Journal of Education in Mathematics, Science and Technology, 2015
Although interview formats support rich data collection in conceptual change studies, interview formats limit sample sizes. This study explores the possibility of using constructed-response formats as an alternative or supplement for collecting similarly rich data across larger pools of subjects in conceptual change studies. While research in…
Descriptors: Interviews, Sample Size, Change, Concept Formation
Malone, Margaret E – Language Testing, 2013
Language assessment literacy refers to language instructors' familiarity with testing definitions and the application of this knowledge to classroom practices in general and specifically to issues related to assessing language. While it is widely agreed that classroom teachers need to assess student progress, many teachers and other test…
Descriptors: Literacy, Language Tests, Interviews, Feedback (Response)
Brantlinger, Andrew; Smith, Beverly – Teachers College Record, 2013
Background/Context: For more than a decade, large alternative teacher certification programs (ATCP) such as the New York City Teaching Fellows (NYCTF) have provided qualified applicants with fast-track or "early entry" routes to paid teaching. While early-entry ATCPs enjoy powerful support in the public and private sectors, critics…
Descriptors: Alternative Teacher Certification, Teacher Education, Preservice Teacher Education, Preservice Teachers
Jones, Ian; Alcock, Lara – Studies in Higher Education, 2014
Peer assessment typically requires students to judge peers' work against assessment criteria. We tested an alternative approach in which students judged pairs of scripts against one another in the absence of assessment criteria. First year mathematics undergraduates (N?=?194) sat a written test on conceptual understanding of multivariable…
Descriptors: Peer Evaluation, Evaluation Criteria, Alternative Assessment, Undergraduate Students
Collier, Lizabeth C. – ProQuest LLC, 2014
This study investigates how university instructors from various disciplines at a large, comprehensive university in the United States evaluate different varieties of English from countries considered "outer circle" (OC) countries, formerly colonized countries where English has been transplanted and is now used unofficially and officially…
Descriptors: Universities, Global Approach, College English, Writing Evaluation
Tokmak, Hatice Sancar; Incikabi, Lutfi; Yelken, Tugba Yanpar – Australasian Journal of Educational Technology, 2012
This comparative case study investigated the educational software evaluation processes of both experts and novices in conjunction with a software evaluation checklist. Twenty novice elementary education students, divided into groups of five, and three experts participated. Each novice group and the three experts evaluated educational software…
Descriptors: Observation, Content Analysis, Focus Groups, Case Studies
Baker, Beverly A. – Assessing Writing, 2010
In high-stakes writing assessments, rater training in the use of a rating scale does not eliminate variability in grade attribution. This realisation has been accompanied by research that explores possible sources of rater variability, such as rater background or rating scale type. However, there has been little consideration thus far of…
Descriptors: Foreign Countries, Writing Evaluation, Writing Tests, Testing
Burmester, Kristen O'Rourke – ProQuest LLC, 2011
Classrooms are a primary site of evidence about learning. Yet classroom proceedings often occur behind closed doors and hence evidence of student learning is observable only to the classroom teacher. The informal and undocumented nature of this information means that it is rarely included in statistical models or quantifiable analyses. This…
Descriptors: Evidence, Student Evaluation, Educational Research, Validity
Hunn, Lorie L. – ProQuest LLC, 2009
This study explored and compared the ways in which school-based cooperating teachers and college supervisors evaluate student teachers. The scores allocated to student teachers by school-based cooperating teachers and college supervisors in the final field experience evaluations of student teachers were analyzed. A mixed methods research design…
Descriptors: Cooperating Teachers, Leadership, Research Design, Student Teachers
Evmenova, Anna S.; Graff, Heidi J.; Jerome, Marci Kinas; Behrmann, Michael M. – Learning Disabilities Research & Practice, 2010
This investigation examined the effects of currently available word prediction software programs that support phonetic/inventive spelling on the quality of journal writing by six students with severe writing and/or spelling difficulties in grades three through six during a month-long summer writing program. A changing conditions single-subject…
Descriptors: Writing Difficulties, Journal Writing, Computer Software Evaluation, Phonetics
Peer reviewedHalleck, Gene B. – Foreign Language Annals, 1996
This study investigated the interrater reliability of proficiency-level judgments of graduate student trainee raters on oral proficiency interviews (OPIs). Trainees' ratings were compared with the judgments of a certified American Council on the Teaching of Foreign Languages (ACTFL) tester for 150 interviews. (Author/JL)
Descriptors: Comparative Analysis, Graduate Students, Higher Education, Interrater Reliability
Previous Page | Next Page »
Pages: 1 | 2
Direct link
