Publication Date
In 2025 | 4 |
Since 2024 | 23 |
Since 2021 (last 5 years) | 47 |
Since 2016 (last 10 years) | 137 |
Since 2006 (last 20 years) | 648 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
Practitioners | 79 |
Researchers | 62 |
Teachers | 40 |
Administrators | 33 |
Policymakers | 27 |
Community | 3 |
Parents | 2 |
Students | 2 |
Counselors | 1 |
Media Staff | 1 |
Location
United Kingdom | 42 |
Australia | 41 |
Canada | 34 |
United Kingdom (England) | 28 |
United States | 25 |
Florida | 19 |
California | 13 |
China | 11 |
Texas | 11 |
Germany | 9 |
Tennessee | 9 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Does not meet standards | 1 |
Rollison, Julia; Hill, Gary; Yu, Ping; Murray, Stephen; Mannix, Danyelle; Mathews-Younes, Anne; Wells, Michael E. – Evaluation and Program Planning, 2012
The Safe Schools/Healthy Students (SS/HS) national evaluation seeks to assess both the implementation process and the results of the SS/HS initiative, exploring factors that have contributed to or detracted from grantee success. Each site is required to forge partnerships with representatives from education, mental health, juvenile justice, and…
Descriptors: School Safety, Child Health, Partnerships in Education, Federal Programs
Ho, Andrew D. – Teachers College Record, 2014
Background/Context: The target of assessment validation is not an assessment but the use of an assessment for a purpose. Although the validation literature often provides examples of assessment purposes, comprehensive reviews of these purposes are rare. Additionally, assessment purposes posed for validation are generally described as discrete and…
Descriptors: Elementary Secondary Education, Standardized Tests, Measurement Objectives, Educational Change
Levin-Rozalis, Miri – Evaluation and Program Planning, 2010
This paper addresses the issue of the knowledge gap between evaluators and the entity being evaluated: the dilemma of the knowledge of professional evaluators vs. the in-depth knowledge of the evaluated subjects. In order to optimize evaluative outcomes, the author suggests an approach based on ideas borrowed from the science of cybernetics as a…
Descriptors: Cybernetics, Evaluators, Knowledge Level, Evaluation Methods
Rink, Judith E. – Research Quarterly for Exercise and Sport, 2013
This article summarizes the research base on teacher effectiveness in physical education from a historical perspective and explores the implications of the recent emphasis on student performance and teacher observation systems to evaluate teachers for physical education. The problems and the potential positive effects of using student performance…
Descriptors: Physical Education, Teacher Effectiveness, Accountability, Teacher Evaluation
Deane, Paul – Assessing Writing, 2013
This paper examines the construct measured by automated essay scoring (AES) systems. AES systems measure features of the text structure, linguistic structure, and conventional print form of essays; as such, the systems primarily measure text production skills. In the current state-of-the-art, AES provide little direct evidence about such matters…
Descriptors: Scoring, Essays, Text Structure, Writing (Composition)
Silvia, Paul J. – Thinking Skills and Creativity, 2011
The present research examined the reliability of three types of divergent thinking tasks (unusual uses, instances, consequences/implications) and two types of subjective scoring (an average across all responses vs. the responses people chose as their top-two responses) within a latent variable framework, using the maximal-reliability "H"…
Descriptors: Scoring, Creative Thinking, Thinking Skills, Test Reliability
Csomay, Eniko; Pollard, Elizabeth; Bordelon, Suzanne; Beck, Audrey – Journal of General Education, 2015
Despite the desire of employers to hire those with the critical-thinking and communication skills a general education (GE) program can offer, the value of GE programs is often questioned due to concerns about four-year graduation rates, perceived low immediate economic payoff, and a dearth of evidence to support their efficacy. This article…
Descriptors: General Education, Critical Thinking, Communication Skills, Graduation Rate
Dombrowski, Stefan C.; Gischlar, Karen L. – Journal of Applied School Psychology, 2014
The authors encourage those in the field of school psychology to consider the use of learning disabilities assessment practices in relation to specific American Psychological Association and National Association of School Psychologists ethical codes and in regard to the American Educational Research Association, American Psychological Association,…
Descriptors: Learning Disabilities, Disability Identification, Ethics, School Psychology
Berliner, David C. – Teacher Educator, 2013
In the United States, but not only here, the movement to evaluate teachers based on student test scores has received powerful political and parental support. The logic is simple. From one testing occasion to another students should show growth in their knowledge and skill. Similar types of students should show similar patterns of growth. Those…
Descriptors: Teacher Evaluation, Merit Pay, Evaluation Problems, Models
Berk, Ronald A. – Journal of Faculty Development, 2013
One of the simplest indicators of teaching or course effectiveness is student ratings on one or more global items from the entire rating scale. That approach seems intuitively sound and easy to use. Global items have even been recommended by a few researchers to get a quick-read, at-a-glance summary for summative decisions about faculty. The…
Descriptors: Rating Scales, Student Evaluation of Teacher Performance, Item Analysis, Test Items
Digby, Joan – Journal of the National Collegiate Honors Council, 2014
Rubric means red ochre--red earth--as in Bryce Canyon and Sedona. Red headers were used in medieval manuscripts as section or chapter markers, and you can bet that the Whore of Babylon got herself some fancy rubrics over the years. Through most of its history, the word has been attached to religious texts and liturgy; rubrics were used as…
Descriptors: Scoring Rubrics, Student Evaluation, Measurement Objectives, Critical Thinking
Rogers, W. Todd – Canadian Journal of Education, 2014
Principals and teachers do not use large-scale assessment results because the lack of distinct and reliable subtests prevents identifying strengths and weaknesses of students and instruction, the results arrive too late to be used, and principals and teachers need assistance to use the results to improve instruction so as to improve student…
Descriptors: Foreign Countries, Group Testing, Multidimensional Scaling, Evaluation Utilization
Stenros, Jaakko; Waern, Annika; Montola, Markus – Simulation & Gaming, 2012
Studying pervasive games is inherently difficult and different from studying computer or board games. This article builds upon the experiences of staging and studying several playful pervasive technology prototypes. It discusses the challenges and pitfalls of evaluating pervasive game prototypes and charts methods that have proven useful in…
Descriptors: Design, Evaluation Methods, Evaluation Problems, Research Methodology
Praetorius, Anna-Katharina; Lenske, Gerlinde; Helmke, Andreas – Learning and Instruction, 2012
Despite considerable interest in the topic of instructional quality in research as well as practice, little is known about the quality of its assessment. Using generalizability analysis as well as content analysis, the present study investigates how reliably and validly instructional quality is measured by observer ratings. Twelve trained raters…
Descriptors: Student Teachers, Interrater Reliability, Content Analysis, Observation
Phillips, Gary W. – Applied Measurement in Education, 2015
This article proposes that sampling design effects have potentially huge unrecognized impacts on the results reported by large-scale district and state assessments in the United States. When design effects are unrecognized and unaccounted for they lead to underestimating the sampling error in item and test statistics. Underestimating the sampling…
Descriptors: State Programs, Sampling, Research Design, Error of Measurement