NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20260
Since 20250
Since 2022 (last 5 years)0
Since 2017 (last 10 years)1
Since 2007 (last 20 years)12
Audience
Assessments and Surveys
What Works Clearinghouse Rating
Showing 1 to 15 of 16 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Roberts, William L.; Boulet, John; Sandella, Jeanne – Advances in Health Sciences Education, 2017
When the safety of the public is at stake, it is particularly relevant for licensing and credentialing exam agencies to use defensible standard setting methods to categorize candidates into competence categories (e.g., pass/fail). The aim of this study was to gather evidence to support change to the Comprehensive Osteopathic Medical Licensing-USA…
Descriptors: Standard Setting, Comparative Analysis, Clinical Experience, Skill Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Baird, Jo-Anne; Black, Paul – Research Papers in Education, 2013
Much has already been written on the controversies surrounding the use of different test theories in educational assessment. Other authors have noted the prevalence of classical test theory over item response theory in practice. This Special Issue draws together articles based upon work conducted on the Reliability Programme for England's…
Descriptors: Test Theory, Foreign Countries, Test Reliability, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Caines, Jade; Engelhard, George, Jr. – Journal of Negro Education, 2012
Standard setting (the process of establishing minimum passing scores on high-stakes exams) is a highly evaluative and policy-driven process. It is a common belief that standard setting panels should be diverse and representative. There is concern, however, that panelists with varying characteristics may differentially influence the results of the…
Descriptors: Geographic Regions, Cutting Scores, Standard Setting, African American Achievement
Peer reviewed Peer reviewed
Direct linkDirect link
Stone, Gregory Ethan; Beltyukova, Svetlana; Fox, Christine M. – International Journal of Testing, 2008
Judge-mediated examinations are defined as those for which expert evaluation (using rubrics) is required to determine correctness, completeness, and reasonability of test-taker responses. The use of multifaceted Rasch modeling has led to improvements in the reliability of scoring such examinations. The establishment of criterion-referenced…
Descriptors: Interrater Reliability, High Stakes Tests, Standard Setting, Minimum Competencies
Peer reviewed Peer reviewed
Direct linkDirect link
Ridley, Charles R.; Shaw-Ridley, Mary – Counseling Psychologist, 2009
Clinical judgment is foundational to psychological practice. Accurate judgment forms the basis for establishing reasonable goals and selecting appropriate treatments, which in turn are essential in achieving positive therapeutic outcomes. Therefore, Spengler and colleagues' meta-analytic finding--clinical judgment accuracy improves marginally with…
Descriptors: Medical Evaluation, Clinical Experience, Inferences, Therapy
Hockett, Jessica A. – Journal for the Education of the Gifted, 2009
Legislative measures designed to ensure that all students meet minimal expectations have concerned leaders in gifted education. In this current educational climate of standards and accountability, however, there is arguably greater agreement than ever before between experts and professional organizations in general education and their counterparts…
Descriptors: Curriculum Design, General Education, Gifted, Educational Indicators
Hardison, Chaitra M.; Vilamovska, Anna-Marie – RAND Corporation, 2009
The Collegiate Learning Assessment (CLA) is a measure of how much students' critical thinking improves after attending college or university. This report illustrates how institutions can set their own standards on the CLA using a method that is appropriate for the CLA's unique characteristics. The authors examined evidence of reliability and…
Descriptors: Standard Setting, Evaluation Methods, Research Reports, Critical Thinking
New York State Education Department, 2014
This technical report provides an overview of the New York State Alternate Assessment (NYSAA), including a description of the purpose of the NYSAA, the processes utilized to develop and implement the NYSAA program, and Stakeholder involvement in those processes. The purpose of this report is to document the technical aspects of the 2013-14 NYSAA.…
Descriptors: Alternative Assessment, Educational Assessment, State Departments of Education, Student Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Black, Beth; Bramley, Tom – Research Papers in Education, 2008
A new judgemental method of equating raw scores on two tests, based on rank-ordering scripts from both tests, has been developed by Bramley. The rank-ordering method has potential application as a judgemental standard-maintaining mechanism, because given a mark on one test (e.g. the A grade boundary mark), the equivalent mark (i.e. at the same…
Descriptors: Foreign Countries, Equated Scores, Test Theory, Evaluative Thinking
Peer reviewed Peer reviewed
Direct linkDirect link
Grainger, Peter; Purnell, Ken; Zipf, Reyna – Assessment & Evaluation in Higher Education, 2008
Decisions by markers about quality in student work remain confusing to most students and markers. This may in part be due to the relatively subjective nature of what constitutes a quality response to an assessment task. This paper reports on an experiment that documented the process of decision-making by multiple markers at a university who…
Descriptors: Student Evaluation, Educational Quality, Achievement Rating, Interrater Reliability
van der Linden, Wim J.; Vos, Hans J.; Chang, Lei – 2000
In judgmental standard setting experiments, it may be difficult to specify subjective probabilities that adequately take the properties of the items into account. As a result, these probabilities are not consistent with each other in the sense that they do not refer to the same borderline level of performance. Methods to check standard setting…
Descriptors: Interrater Reliability, Judges, Probability, Standard Setting
Hammersley, Martyn – International Journal of Research & Method in Education, 2007
This article addresses the perennial issue of the criteria by which qualitative research should be evaluated. At the present time, there is a sharp conflict between demands for explicit criteria, for example in order to serve systematic reviewing and evidence-based practice, and arguments on the part of some qualitative researchers that such…
Descriptors: Qualitative Research, Research Methodology, Evaluation Criteria, Research Problems
Hertz, Norman R.; Chinn, Roberta N. – 2002
Nearly all of the research on standard setting focuses on different standard setting methods rather than the interaction of group members and the instructions given to group members. This study explored the effect of deliberation style and the requirement to reach consensus on the passing score, on rater satisfaction, and on postdecision…
Descriptors: Decision Making, Evaluation Methods, Evaluators, Interaction
Peer reviewed Peer reviewed
Direct linkDirect link
Meyen, Edward; Bui, Yvonne N. – Journal of Technology and Teacher Education, 2003
The Online Academy (HO29K73002) was funded by the Office of Special Education Programs (OSEP) to develop research-based online instructional modules in the content areas of reading, positive behavior support and technology across the curriculum. Targeted to preservice teacher education programs in Institutions of Higher Education (IHE), but also…
Descriptors: Teacher Education Programs, Learning Modules, Program Descriptions, Online Systems
New Mexico Public Education Department, 2007
The purpose of the NMSBA technical report is to provide users and other interested parties with a general overview of and technical characteristics of the 2007 NMSBA. The 2007 technical report contains the following information: (1) Test development; (2) Scoring procedures; (3) Summary of student performance; (4) Statistical analyses of item and…
Descriptors: Interrater Reliability, Standard Setting, Measures (Individuals), Scoring
Previous Page | Next Page ยป
Pages: 1  |  2