NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Showing 1 to 15 of 16 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Andrew P. Jaciw – American Journal of Evaluation, 2025
By design, randomized experiments (XPs) rule out bias from confounded selection of participants into conditions. Quasi-experiments (QEs) are often considered second-best because they do not share this benefit. However, when results from XPs are used to generalize causal impacts, the benefit from unconfounded selection into conditions may be offset…
Descriptors: Elementary School Students, Elementary School Teachers, Generalization, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Pamela R. Buckley; Katie Massey Combs; Karen M. Drewelow; Brittany L. Hubler; Marion Amanda Lain – Evaluation Review, 2025
As evidence-based interventions are scaled, fidelity of implementation, and thus effectiveness, often wanes. Validated fidelity measures can improve researchers' ability to attribute outcomes to the intervention and help practitioners feel more confident in implementing the intervention as intended. We aim to provide a model for the validation of…
Descriptors: Middle School Students, Middle School Teachers, Evidence Based Practice, Program Development
Peer reviewed Peer reviewed
Direct linkDirect link
Jieun Kim; Daniel Richard Isbell – Language Assessment Quarterly, 2024
The ACTFL Assessment of Performance Toward Proficiency in Languages (AAPPL, https://www.actfl.n.d.org/assessments/k-12-assessments/aappl) assesses proficiency in 11 languages for students in grades 3 to 12 and is often used to award the Seal of Biliteracy. While arguments for the valid interpretation and uses of the AAPPL have previously been…
Descriptors: Language Tests, Second Language Learning, Second Language Instruction, Language Proficiency
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Shijun, Chen – International Education Studies, 2022
The high-stakes College English Test (CET), developed, administered, and reformed over the last 20 years, has received great attention in the aspect of washback on teaching and learning from previous research. Very few studies explored its consequences in the workplace domain--being used as a screening lever. This research aimed to 1) compare…
Descriptors: Language Tests, Test Use, Second Language Learning, Second Language Instruction
Peer reviewed Peer reviewed
Direct linkDirect link
Jin, Hui; van Rijn, Peter; Moore, John C.; Bauer, Malcolm I.; Pressler, Yamina; Yestness, Nissa – International Journal of Science Education, 2019
This article provides a validation framework for research on the development and use of science Learning Progressions (LPs). The framework describes how evidence from various sources can be used to establish an interpretive argument and a validity argument at five stages of LP research--development, scoring, generalisation, extrapolation, and use.…
Descriptors: Sequential Approach, Educational Research, Science Education, Validity
Jin, Hui; van Rijn, Peter; Moore, John C.; Bauer, Malcolm I.; Pressler, Yamina; Yestness, Nissa – Grantee Submission, 2019
This article provides a validation framework for research on the development and use of science Learning Progressions (LPs). The framework describes how evidence from various sources can be used to establish an interpretive argument and a validity argument at five stages of LP research--development, scoring, generalisation, extrapolation, and use.…
Descriptors: Sequential Approach, Educational Research, Science Education, Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Ketterlin-Geller, Leanne R.; Perry, Lindsey; Adams, Elizabeth – Applied Measurement in Education, 2019
Despite the call for an argument-based approach to validity over 25 years ago, few examples exist in the published literature. One possible explanation for this outcome is that the complexity of the argument-based approach makes implementation difficult. To counter this claim, we propose that the Assessment Triangle can serve as the overarching…
Descriptors: Validity, Educational Assessment, Models, Screening Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Brennan, Robert L. – Journal of Educational Measurement, 2013
Kane's paper "Validating the Interpretations and Uses of Test Scores" is the most complete and clearest discussion yet available of the argument-based approach to validation. At its most basic level, validation as formulated by Kane is fundamentally a simply-stated two-step enterprise: (1) specify the claims inherent in a particular interpretation…
Descriptors: Validity, Test Interpretation, Test Use, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Hatala, Rose; Cook, David A.; Brydges, Ryan; Hawkins, Richard – Advances in Health Sciences Education, 2015
In order to construct and evaluate the validity argument for the Objective Structured Assessment of Technical Skills (OSATS), based on Kane's framework, we conducted a systematic review. We searched MEDLINE, EMBASE, CINAHL, PsycINFO, ERIC, Web of Science, Scopus, and selected reference lists through February 2013. Working in duplicate, we selected…
Descriptors: Measures (Individuals), Test Validity, Surgery, Skills
Peer reviewed Peer reviewed
Direct linkDirect link
Kane, Michael T. – Journal of Educational Measurement, 2013
This response to the comments contains three main sections, each addressing a subset of the comments. In the first section, I will respond to the comments by Brennan, Haertel, and Moss. All of these comments suggest ways in which my presentation could be extended or improved; I generally agree with their suggestions, so my response to their…
Descriptors: Validity, Test Interpretation, Test Use, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Kane, Michael T. – Journal of Educational Measurement, 2013
To validate an interpretation or use of test scores is to evaluate the plausibility of the claims based on the scores. An argument-based approach to validation suggests that the claims based on the test scores be outlined as an argument that specifies the inferences and supporting assumptions needed to get from test responses to score-based…
Descriptors: Test Interpretation, Validity, Scores, Test Use
Jia, Yujie – ProQuest LLC, 2013
This study employed Bachman and Palmer's (2010) Assessment Use Argument framework to investigate to what extent the use of a second language oral test as an exit test in a Hong Kong university can be justified. It also aimed to help test developers of this oral test identify the most critical areas in the current test design that might need…
Descriptors: Test Use, Language Tests, Oral Language, Second Language Learning
Wang, Huan – ProQuest LLC, 2010
Multiple uses of the same assessment may present challenges for both the design and use of an assessment. Little advice, however, has been given to assessment developers as to how to understand the phenomena of multiple assessment use and meet the challenges these present. Particularly problematic is the case in which an assessment is used for…
Descriptors: Test Use, Testing Programs, Program Effectiveness, Test Construction
Peer reviewed Peer reviewed
Denner, Peter R.; Salzman, Stephanie A.; Bangert, Arthur W. – Journal of Personnel Evaluation in Education, 2001
Examined the validity and generalizability of the use of Teacher Work Samples to assess the ability of preservice teachers and inservice teachers to meet national and state teaching standards and to make an impact on the learning of their students. Results of the study, which involved 132 work samples, show initial support for teacher work sample…
Descriptors: Academic Achievement, Elementary Secondary Education, Generalization, Preservice Teachers
Peer reviewed Peer reviewed
Dunbar, Stephen B.; And Others – Applied Measurement in Education, 1991
Issues pertaining to the quality of performance assessments, including reliability and validity, are discussed. The relatively limited generalizability of performance across tasks is indicative of the care needed to evaluate performance assessments. Quality control is an empirical matter when measurement is intended to inform public policy. (SLD)
Descriptors: Educational Assessment, Generalization, Interrater Reliability, Measurement Techniques
Previous Page | Next Page »
Pages: 1  |  2