NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Practitioners1
What Works Clearinghouse Rating
Showing 1 to 15 of 72 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Russell, Michael – Educational Measurement: Issues and Practice, 2022
Despite agreement about the central importance of validity for educational and psychological testing, consensus regarding the definition of validity remains elusive. Differences in the definition of validity are examined and reveals that a potential cause of disagreement stems from differences in word use and meanings given to key terms commonly…
Descriptors: Test Validity, Psychological Testing, Educational Testing, Vocabulary
Peer reviewed Peer reviewed
Direct linkDirect link
Kylie Gorney; Sandip Sinharay – Educational and Psychological Measurement, 2025
Test-takers, policymakers, teachers, and institutions are increasingly demanding that testing programs provide more detailed feedback regarding test performance. As a result, there has been a growing interest in the reporting of subscores that potentially provide such detailed feedback. Haberman developed a method based on classical test theory…
Descriptors: Scores, Test Theory, Test Items, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Hogan, Thomas; DeStefano, Marissa; Gilby, Caitlin; Kosman, Dana; Peri, Joshua – Applied Measurement in Education, 2021
Buros' "Mental Measurements Yearbook (MMY)" has provided professional reviews of commercially published psychological and educational tests for over 80 years. It serves as a kind of conscience for the testing industry. For a random sample of 50 entries in the "19th MMY" (a total of 100 separate reviews) this study determined…
Descriptors: Test Reviews, Interrater Reliability, Psychological Testing, Educational Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Suthathip Thirakunkovit – Language Testing in Asia, 2025
Establishing a cut score is a crucial aspect of the test development process since the selected cut score has the potential to impact students' performance outcomes and shape instructional strategies within the classroom. Therefore, it is vital for those involved in test development to set a cut score that is both fair and justifiable. This cut…
Descriptors: Cutting Scores, Culture Fair Tests, Language Tests, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Hong, Seong Eun; Monroe, Scott; Falk, Carl F. – Journal of Educational Measurement, 2020
In educational and psychological measurement, a person-fit statistic (PFS) is designed to identify aberrant response patterns. For parametric PFSs, valid inference depends on several assumptions, one of which is that the item response theory (IRT) model is correctly specified. Previous studies have used empirical data sets to explore the effects…
Descriptors: Educational Testing, Psychological Testing, Goodness of Fit, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Nisbet, Isabel; Shaw, Stuart D. – Assessment in Education: Principles, Policy & Practice, 2019
Fairness in assessment is seen as increasingly important but there is a need for greater clarity in use of the term 'fair'. Also, fairness is perceived through a range of 'lenses' reflecting different traditions of thought. The lens used determines how fairness is seen and described. This article distinguishes different uses of 'fair' which have…
Descriptors: Test Bias, Measurement, Theories, Educational Assessment
Sinharay, Sandip; van Rijn, Peter W. – Journal of Educational and Behavioral Statistics, 2020
Response time models (RTMs) are of increasing interest in educational and psychological testing. This article focuses on the lognormal model for response times, which is one of the most popular RTMs. Several existing statistics for testing normality and the fit of factor analysis models are repurposed for testing the fit of the lognormal model. A…
Descriptors: Educational Testing, Psychological Testing, Goodness of Fit, Factor Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Beck, Klaus – Frontline Learning Research, 2020
Many test developers try to ensure the content validity of their tests by having external experts review the items, e.g. in terms of relevance, difficulty, or clarity. Although this approach is widely accepted, a closer look reveals several pitfalls need to be avoided if experts' advice is to be truly helpful. The purpose of this paper is to…
Descriptors: Content Validity, Psychological Testing, Educational Testing, Student Evaluation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Mousavi, Amin; Cui, Ying – Education Sciences, 2020
Often, important decisions regarding accountability and placement of students in performance categories are made on the basis of test scores generated from tests, therefore, it is important to evaluate the validity of the inferences derived from test results. One of the threats to the validity of such inferences is aberrant responding. Several…
Descriptors: Student Evaluation, Educational Testing, Psychological Testing, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip – Journal of Educational Measurement, 2018
Response-time models are of increasing interest in educational and psychological testing. This article focuses on the lognormal model for response times, which is one of the most popular response-time models, and suggests a simple person-fit statistic for the model. The distribution of the statistic under the null hypothesis of no misfit is proved…
Descriptors: Reaction Time, Educational Testing, Psychological Testing, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Jonson, Jessica L.; Trantham, Pamela; Usher-Tate, Betty Jean – Educational Measurement: Issues and Practice, 2019
One of the substantive changes in the 2014 Standards for Educational and Psychological Testing was the elevation of fairness in testing as a foundational element of practice in addition to validity and reliability. Previous research indicates that testing practices often do not align with professional standards and guidelines. Therefore, to raise…
Descriptors: Culture Fair Tests, Test Validity, Test Reliability, Intelligence Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Plake, Barbara S.; Wise, Lauress L. – Educational Measurement: Issues and Practice, 2014
With the 2014 publication of the 5th revision of the "Standards for Educational and Psychological Testing," the cochairs of the Joint Committee for the revision process were asked to consider the role and importance of the "Standards" for the educational testing community, and in particular for members of the National Council…
Descriptors: Standards, Educational Testing, Psychological Testing, Role
Peer reviewed Peer reviewed
Direct linkDirect link
Newton, Paul E.; Shaw, Stuart D. – Assessment in Education: Principles, Policy & Practice, 2016
The ability to convey shared meaning with minimal ambiguity is highly desirable for technical terms within disciplines and professions. Unfortunately, there is no widespread professional consensus over the meaning of the word "validity" as it pertains to educational and psychological testing. After illustrating the nature and extent of…
Descriptors: Test Validity, Validity, Ambiguity (Semantics), Psychological Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Sireci, Stephen G. – Assessment in Education: Principles, Policy & Practice, 2016
A misconception exists that validity may refer only to the "interpretation" of test scores and not to the "uses" of those scores. The development and evolution of validity theory illustrate test score interpretation was a primary focus in the earliest days of modern testing, and that validating interpretations derived from test…
Descriptors: Test Validity, Misconceptions, Evaluation Utilization, Data Interpretation
Peer reviewed Peer reviewed
Direct linkDirect link
Veldkamp, Bernard P. – Journal of Educational Measurement, 2016
Many standardized tests are now administered via computer rather than paper-and-pencil format. The computer-based delivery mode brings with it certain advantages. One advantage is the ability to adapt the difficulty level of the test to the ability level of the test taker in what has been termed computerized adaptive testing (CAT). A second…
Descriptors: Computer Assisted Testing, Reaction Time, Standardized Tests, Difficulty Level
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5