NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 10 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Suthathip Thirakunkovit – Language Testing in Asia, 2025
Establishing a cut score is a crucial aspect of the test development process since the selected cut score has the potential to impact students' performance outcomes and shape instructional strategies within the classroom. Therefore, it is vital for those involved in test development to set a cut score that is both fair and justifiable. This cut…
Descriptors: Cutting Scores, Culture Fair Tests, Language Tests, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Clauser, Jerome C.; Hambleton, Ronald K.; Baldwin, Peter – Educational and Psychological Measurement, 2017
The Angoff standard setting method relies on content experts to review exam items and make judgments about the performance of the minimally proficient examinee. Unfortunately, at times content experts may have gaps in their understanding of specific exam content. These gaps are particularly likely to occur when the content domain is broad and/or…
Descriptors: Scores, Item Analysis, Classification, Decision Making
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Papageorgiou, Spiros; Wu, Sha; Hsieh, Ching-Ni; Tannenbaum, Richard J.; Cheng, Mengmeng – ETS Research Report Series, 2019
The past decade has seen an emerging interest in mapping (aligning or linking) test scores to language proficiency levels of external performance scales or frameworks, such as the Common European Framework of Reference (CEFR), as well as locally developed frameworks, such as China's Standards of English Language Ability (CSE). Such alignment is…
Descriptors: English (Second Language), Language Tests, Second Language Learning, Computer Assisted Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Skaggs, Gary; Hein, Serge F.; Wilkins, Jesse L. M. – Journal of Educational Measurement, 2016
This article introduces the Diagnostic Profiles (DP) standard setting method for setting a performance standard on a test developed from a cognitive diagnostic model (CDM), the outcome of which is a profile of mastered and not-mastered skills or attributes rather than a single test score. In the DP method, the key judgment task for panelists is a…
Descriptors: Models, Standard Setting, Profiles, Diagnostic Tests
Peer reviewed Peer reviewed
Norcini, John J.; Shea, Judy A. – Applied Measurement in Education, 1997
The major forms of evidence that support a standard's credibility are reviewed, and what can be done over time and for different forms of an examination to enhance its comparability in a credentialing setting is outlined. Pass-fail decisions must be consistent to ensure a standard's credibility. (SLD)
Descriptors: Certification, Comparative Analysis, Credentials, Credibility
Peer reviewed Peer reviewed
Noah, Harold J. – Comparative Education Review, 1984
Comparative education can deepen understanding of our own education and society, be of assistance to policymakers and administrators, and form a most valuable part of teacher education. Like all applied fields, however, it is open to potential abuse by those who would use its results to support--or oppose--specific change programs. (BRR)
Descriptors: Adoption (Ideas), Comparative Analysis, Comparative Education, Decision Making
Poggio, John P.; Glasnapp, Douglas R. – 1994
This paper reports on a newly designed judgmental method for setting test performance standard that: (1) overcome many of the practical and psychometric problems associated with the Angoff and Ebel methods; (2) can be used to set multiple cut points on a score scale; (3) may be readily and efficiently implemented with assessments that use…
Descriptors: Comparative Analysis, Constructed Response, Cutting Scores, Decision Making
Peer reviewed Peer reviewed
Norcini, John J. – Journal of Educational Measurement, 1990
Whether cutting score equivalents (CSEs) based on examinee performance are the same as CSEs based on expert judgment was examined using data from 3,262 examinees taking an internal medicine certification examination. CSEs produced by 40 physicians/experts were closer to the criteria than were standards derived from examinee performance. (SLD)
Descriptors: Certification, Comparative Analysis, Cutting Scores, Decision Making
Peer reviewed Peer reviewed
Plake, Barbara S. – Applied Measurement in Education, 1995
This article provides a framework for the rest of the articles in this special issue comparing the utility of three standard-setting methods with complex performance assessments. The context of the standard setting study is described, and the methods are outlined. (SLD)
Descriptors: Comparative Analysis, Criteria, Decision Making, Educational Assessment
Peer reviewed Peer reviewed
Woehr, David J.; And Others – Educational and Psychological Measurement, 1991
Methods for setting cutoff scores based on criterion performance, normative comparison, and absolute judgment were compared for scores on a multiple-choice psychology examination for 121 undergraduates and 251 undergraduates as a comparison group. All methods fell within the standard error of measurement. Implications of differences for decision…
Descriptors: Comparative Analysis, Concurrent Validity, Content Validity, Cutting Scores