NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20260
Since 20250
Since 2022 (last 5 years)0
Since 2017 (last 10 years)0
Since 2007 (last 20 years)9
What Works Clearinghouse Rating
Showing 1 to 15 of 198 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Zumbo, Bruno D.; Hubley, Anita M. – Assessment in Education: Principles, Policy & Practice, 2016
Ultimately, measures in research, testing, assessment and evaluation are used, or have implications, for ranking, intervention, feedback, decision-making or policy purposes. Explicit recognition of this fact brings the often-ignored and sometimes maligned concept of consequences to the fore. Given that measures have personal and social…
Descriptors: Testing Programs, Testing Problems, Measurement Techniques, Student Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Koretz, Daniel – Assessment in Education: Principles, Policy & Practice, 2016
Daniel Koretz is the Henry Lee Shattuck Professor of Education at the Harvard Graduate School of Education. His research focuses on educational assessment and policy, particularly the effects of high-stakes testing on educational practice and the validity of score gains. He is the author of "Measuring Up: What Educational Testing Really Tells…
Descriptors: Test Validity, Definitions, Evidence, Relevance (Education)
Peer reviewed Peer reviewed
Direct linkDirect link
Koretz, Daniel – Measurement: Interdisciplinary Research and Perspectives, 2013
Haertel's argument is that one must "expand the scope of test validation to include indirect testing effects" because these effects are often the "rationale for the entire testing program." The author strongly agrees that this is essential. However, he maintains that Haertel's argument does not go far enough and that there are two additional…
Descriptors: Educational Testing, Test Validity, Test Results, Testing Programs
Peer reviewed Peer reviewed
Direct linkDirect link
Lane, Suzanne – Measurement: Interdisciplinary Research and Perspectives, 2012
Considering consequences in the evaluation of validity is not new although it is still debated by Paul E. Newton and others. The argument-based approach to validity entails an interpretative argument that explicitly identifies the proposed interpretations and uses of test scores and a validity argument that provides a structure for evaluating the…
Descriptors: Educational Opportunities, Accountability, Validity, Inferences
Peer reviewed Peer reviewed
Direct linkDirect link
Starr, Karen – AASA Journal of Scholarship & Practice, 2014
This article is a commentary on Australia's involvement in the Programme for International Student Assessment (PISA) tests. It provides a rationale for Australia's participation in the PISA programme, the influences of PISA involvement on education policies and practices, and considerations and implications for school leaders and education…
Descriptors: Foreign Countries, Educational Policy, Educational Practices, Comparative Education
Peer reviewed Peer reviewed
Direct linkDirect link
Allen, Martin – FORUM: for promoting 3-19 comprehensive education, 2013
Well before the examinations grade crisis of 2012, Michael Gove had set out clear intentions for reforming public examinations. Though he claimed to be improving examinations and assessment by replicating practices that took place in high-performing countries and thus improving the ability of the UK economy to "compete", this…
Descriptors: Foreign Countries, Academic Standards, Educational Change, Standardized Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Wagner, Daniel A.; Lockheed, Marlaine; Mullis, Ina; Martin, Michael O.; Kanjee, Anil; Gove, Amber; Dowd, Amy Jo – Compare: A Journal of Comparative and International Education, 2012
Over the past decade, international and national education agencies have begun to emphasize the improvement of the quality (rather than quantity) of education in developing countries. This trend has been paralleled by a significant increase in the use of educational assessments as a way to measure gains and losses in quality of learning. As…
Descriptors: Developing Nations, Foreign Countries, Educational Assessment, Reading Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Ferrara, Steve; Perie, Marianne; Johnson, Eugene – Journal of Applied Testing Technology, 2008
Psychometricians continue to introduce new approaches to setting cut scores for educational assessments in an attempt to improve on current methods. In this paper we describe the Item-Descriptor (ID) Matching method, a method based on IRT item mapping. In ID Matching, test content area experts match items (i.e., their judgments about the knowledge…
Descriptors: Test Results, Test Content, Testing Programs, Educational Testing
Green, Donald Ross – 1997
It is argued that publishers of achievement tests, especially those who publish tests intended for use in many parts of the United States, are for the most part not in a position to obtain any decent evidence about the consequences of the uses that are made of their tests. What responsibilities and actions publishers can reasonably be expected to…
Descriptors: Achievement Tests, Standardized Tests, State Programs, Test Use
Green, Donald Ross – 1985
The use of item banks and item response theory has resulted in new ways to misinterpret and misuse tests through customized, yet standardized, achievement test batteries. The new test batteries create the possibility of serious misunderstandings based on the idea that any subset of items from the pool with a proper range of difficulties will…
Descriptors: Academic Achievement, Achievement Gains, Item Banks, Latent Trait Theory
Noggle, Nelson L. – 1979
The potential use of criterion referenced tests (CRT) and norm referenced tests (NRT) in the same testing program is discussed. The advantages and disadvantages of each are listed, and the best solution, a merging, is proposed. To merge CRTs and NRTs into an overall testing program, meaningful and useful to all levels, consideration must be given…
Descriptors: Achievement Tests, Computer Assisted Testing, Criterion Referenced Tests, Diagnostic Tests
Arter, Judith A. – 1982
Specific recommendations are made concerning the circumstances under which the benefits of out-of-level testing outweigh the problems associated with it. Topics explored are: various methods for deciding when a set of test scores is invalid and the utility of these methods for local evaluators, the accuracy of vertical scaling, and the usefulness…
Descriptors: Equated Scores, Evaluation Methods, Local Norms, Scores
Peer reviewed Peer reviewed
Mercer, Jane R. – School Psychology Digest, 1979
A defense of the System of Multicultural Pluralistic Assessment (SOMPA) in response to previous articles in this issue of this journal is presented by the test developer. The defense is detailed and addresses each previous article. (JKS)
Descriptors: Culture Fair Tests, Educational Testing, Elementary Secondary Education, Exceptional Persons
Herman, Joan L. – 1986
Issues in designing valid tests for the National Assessment of Educational Progress (NAEP) are discussed. Test scores are often provided without any information on the nature of the tasks represented. Because test domains are defined by individual item writers, the generalizability between tests and items is suspect. While typical content…
Descriptors: Achievement Tests, Content Validity, Criterion Referenced Tests, Educational Assessment
Baker, Keith – Phi Delta Kappan, 2007
The idea that America was being harmed because its schools were not keeping up with those in other advanced nations emerged after Sputnik in 1957, took a firm hold on education policy when "A Nation at Risk" appeared in 1983, and continues today. Policy makers justify this concern by pointing to evidence showing that, for individuals…
Descriptors: Testing Programs, Academic Achievement, Achievement Tests, International Education
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  14