Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 3 |
| Since 2017 (last 10 years) | 3 |
| Since 2007 (last 20 years) | 8 |
Descriptor
| Evaluation Methods | 12 |
| Validity | 12 |
| Program Evaluation | 5 |
| Measurement | 4 |
| Test Construction | 4 |
| Test Use | 4 |
| Evaluation Utilization | 3 |
| Reliability | 3 |
| Diagnostic Tests | 2 |
| Educational Assessment | 2 |
| Educational Improvement | 2 |
| More ▼ | |
Source
| Educational Measurement:… | 12 |
Author
| Abedi, Jamal | 1 |
| Alina A. von Davier | 1 |
| Burkett, Ruth S. | 1 |
| Burling, Kelly S. | 1 |
| Deborah J. Harris | 1 |
| Downing, Joyce Anderson | 1 |
| Herman, Joan L. | 1 |
| Ji, Xuejun Ryan | 1 |
| Jiangang Hao | 1 |
| Klein, Davina C. D. | 1 |
| Lamson, Sharon | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 12 |
| Reports - Descriptive | 5 |
| Reports - Research | 4 |
| Reports - Evaluative | 3 |
| Information Analyses | 1 |
| Opinion Papers | 1 |
Education Level
| Elementary Education | 1 |
| Elementary Secondary Education | 1 |
| Grade 4 | 1 |
| Higher Education | 1 |
| Intermediate Grades | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
| ACT Assessment | 1 |
| Progress in International… | 1 |
What Works Clearinghouse Rating
Jiangang Hao; Alina A. von Davier; Victoria Yaneva; Susan Lottridge; Matthias von Davier; Deborah J. Harris – Educational Measurement: Issues and Practice, 2024
The remarkable strides in artificial intelligence (AI), exemplified by ChatGPT, have unveiled a wealth of opportunities and challenges in assessment. Applying cutting-edge large language models (LLMs) and generative AI to assessment holds great promise in boosting efficiency, mitigating bias, and facilitating customized evaluations. Conversely,…
Descriptors: Evaluation Methods, Artificial Intelligence, Educational Change, Computer Software
Lewis, Jennifer; Lim, Hwanggyu; Padellaro, Frank; Sireci, Stephen G.; Zenisky, April L. – Educational Measurement: Issues and Practice, 2022
Setting cut scores on (MSTs) is difficult, particularly when the test spans several grade levels, and the selection of items from MST panels must reflect the operational test specifications. In this study, we describe, illustrate, and evaluate three methods for mapping panelists' Angoff ratings into cut scores on the scale underlying an MST. The…
Descriptors: Cutting Scores, Adaptive Testing, Test Items, Item Analysis
Ji, Xuejun Ryan; Wu, Amery D. – Educational Measurement: Issues and Practice, 2023
The Cross-Classified Mixed Effects Model (CCMEM) has been demonstrated to be a flexible framework for evaluating reliability by measurement specialists. Reliability can be estimated based on the variance components of the test scores. Built upon their accomplishment, this study extends the CCMEM to be used for evaluating validity evidence.…
Descriptors: Measurement, Validity, Reliability, Models
Suto, Irenka – Educational Measurement: Issues and Practice, 2012
Internationally, many assessment systems rely predominantly on human raters to score examinations. Arguably, this facilitates the assessment of multiple sophisticated educational constructs, strengthening assessment validity. It can introduce subjectivity into the scoring process, however, engendering threats to accuracy. The present objectives…
Descriptors: Evaluation Methods, Scoring, Qualitative Research, Protocol Analysis
Parkes, Jay – Educational Measurement: Issues and Practice, 2007
Reliability consists of both important social and scientific values and methods for evidencing those values, though in practice methods are often conflated with the values. With the two distinctly understood, a reliability argument can be made that articulates the particular reliability values most relevant to the particular measurement situation…
Descriptors: Validity, Reliability, Evaluation Methods, Measurement
Yao, Yuankun; Thomas, Matt; Nickens, Nicole; Downing, Joyce Anderson; Burkett, Ruth S.; Lamson, Sharon – Educational Measurement: Issues and Practice, 2008
This study applied Messick's unified, multifaceted concept of construct validity to an electronic portfolio system used in a teacher education program. The subjects included 128 preservice teachers who recently completed their final portfolio reviews and student teaching experiences. Four of Messick's six facets of validity were investigated for…
Descriptors: Student Teaching, Portfolios (Background Materials), Preservice Teachers, Preservice Teacher Education
Nichols, Paul D.; Meyers, Jason L.; Burling, Kelly S. – Educational Measurement: Issues and Practice, 2009
Assessments labeled as formative have been offered as a means to improve student achievement. But labels can be a powerful way to miscommunicate. For an assessment use to be appropriately labeled "formative," both empirical evidence and reasoned arguments must be offered to support the claim that improvements in student achievement can be linked…
Descriptors: Academic Achievement, Tutoring, Student Evaluation, Evaluation Methods
Peer reviewedReckase, Mark D. – Educational Measurement: Issues and Practice, 1998
Considers what a responsible test developer would do to gain information to support the consequential basis of validity for a test early in the development. How the consequential basis of validity of the program would be monitored and reported during the life of the program is examined. The validity of the ACT Assessment is considered as if it…
Descriptors: Evaluation Methods, Program Evaluation, Test Construction, Validity
Peer reviewedYen, Wendy M. – Educational Measurement: Issues and Practice, 1998
The articles in this issue, written from the perspectives of academics, practitioners, and publishers, show that examining the consequences of assessment is an important, large, and difficult task. Collaborative action by assessment developers, users, and the educational measurement community is needed if progress is to be made. (SLD)
Descriptors: Cooperation, Evaluation Methods, Program Evaluation, Responsibility
Peer reviewedMoss, Pamela A. – Educational Measurement: Issues and Practice, 1998
Provides an argument for incorporating consideration of consequences into validity theory that is grounded in the reflexive nature of social knowledge. It also calls for the consideration of evidence of validity based on the actual discourse surrounding the practices and products of testing. (SLD)
Descriptors: Evaluation Methods, Evaluation Utilization, Program Evaluation, Test Construction
Shepard, Lorrie A. – Educational Measurement: Issues and Practice, 2009
In many school districts, the pressure to raise test scores has created overnight celebrity status for formative assessment. Its powers to raise student achievement have been touted, however, without attending to the research on which these claims were based. Sociocultural learning theory provides theoretical grounding for understanding how…
Descriptors: Learning Theories, Validity, Student Evaluation, Evaluation Methods
Peer reviewedHerman, Joan L.; Klein, Davina C. D.; Abedi, Jamal – Educational Measurement: Issues and Practice, 2000
Explores methods of assessing opportunity to learn (OTL) and presents data collected as part of a pilot study of an eighth-grade statewide mathematics assessment to explore issues of validity. Investigates the integrity of various dimensions thought to constitute OTL, analyzes the relationships among teachers' and students' self-reports, and draws…
Descriptors: Educational Policy, Elementary Secondary Education, Evaluation Methods, Junior High School Students

Direct link
