NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1 to 15 of 75 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Andrew P. Jaciw – American Journal of Evaluation, 2025
By design, randomized experiments (XPs) rule out bias from confounded selection of participants into conditions. Quasi-experiments (QEs) are often considered second-best because they do not share this benefit. However, when results from XPs are used to generalize causal impacts, the benefit from unconfounded selection into conditions may be offset…
Descriptors: Elementary School Students, Elementary School Teachers, Generalization, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Liou, Gloria; Bonner, Cavan V.; Tay, Louis – International Journal of Testing, 2022
With the advent of big data and advances in technology, psychological assessments have become increasingly sophisticated and complex. Nevertheless, traditional psychometric issues concerning the validity, reliability, and measurement bias of such assessments remain fundamental in determining whether score inferences of human attributes are…
Descriptors: Psychometrics, Computer Assisted Testing, Adaptive Testing, Data
Peer reviewed Peer reviewed
Direct linkDirect link
Wesolowski, Brian C. – Music Educators Journal, 2020
Validity, reliability, and fairness are three prominent indicators for evaluating the quality of assessment processes. Each of the indicators is most often written about and applied in the context of large-scale assessment. As a result, the technical properties of these indicators make them limited in both their practicality and relevance for…
Descriptors: Music Education, Test Validity, Test Reliability, Student Evaluation
Meyer, J. Patrick; Dahlin, Michael – NWEA, 2022
The MAP® Growth™ theory of action describes key features of MAP Growth and its position in a comprehensive assessment system. The basic premise of the theory of action is that all students learn when MAP Growth is situated in a comprehensive assessment system and used for its intended purposes to yield information about student learning and enable…
Descriptors: Achievement Tests, Academic Achievement, Achievement Gains, Student Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Angela Johnson; Elizabeth Barker; Marcos Viveros Cespedes – Educational Measurement: Issues and Practice, 2024
Educators and researchers strive to build policies and practices on data and evidence, especially on academic achievement scores. When assessment scores are inaccurate for specific student populations or when scores are inappropriately used, even data-driven decisions will be misinformed. To maximize the impact of the research-practice-policy…
Descriptors: Equal Education, Inclusion, Evaluation Methods, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Leighton, Jacqueline P.; Lehman, Blair – Educational Measurement: Issues and Practice, 2020
In this digital ITEMS module, Dr. Jacqueline Leighton and Dr. Blair Lehman review differences between think-aloud interviews to measure problem-solving processes and cognitive labs to measure comprehension processes. Learners are introduced to historical, theoretical, and procedural differences between these methods and how to use and analyze…
Descriptors: Protocol Analysis, Interviews, Problem Solving, Cognitive Processes
Flanagan, Agnes; Cormier, Damien C. – Communique, 2019
One of the areas subsumed under the data-based decision making and accountability practice identified in the National Association of School Psychologists' (NASP) "Model for Integrated School Psychological Services" is to collect information on psychological and educational variables to make decisions at a number of levels of service…
Descriptors: Test Bias, School Psychologists, Measurement, Data Collection
Center on Standards and Assessments Implementation, 2018
Reliability is a measure of consistency. It is the degree to which student results are the same when they take the same test on different occasions, when different scorers score the same item or task, and when different but equivalent tests are taken at the same time or at different times. Reliability is about making sure that different test forms…
Descriptors: Test Reliability, Test Validity, Student Evaluation, Test Bias
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Schmidgall, Jonathan; Cid, Jaime; Carter Grissom, Elizabeth; Li, Lucy – ETS Research Report Series, 2021
The redesigned "TOEIC Bridge"® tests were designed to evaluate test takers' English listening, reading, speaking, and writing skills in the context of everyday adult life. In this paper, we summarize the initial validity argument that supports the use of test scores for the purpose of selection, placement, and evaluation of a test…
Descriptors: Language Tests, Second Language Learning, English (Second Language), Language Proficiency
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Baghaei, Purya; Kubinger, Klaus D. – Practical Assessment, Research & Evaluation, 2015
The present paper gives a general introduction to the linear logistic test model (Fischer, 1973), an extension of the Rasch model with linear constraints on item parameters, along with eRm (an R package to estimate different types of Rasch models; Mair, Hatzinger, & Mair, 2014) functions to estimate the model and interpret its parameters. The…
Descriptors: Item Response Theory, Models, Test Validity, Hypothesis Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Huggins, Anne C.; Penfield, Randall D. – Educational Measurement: Issues and Practice, 2012
A goal for any linking or equating of two or more tests is that the linking function be invariant to the population used in conducting the linking or equating. Violations of population invariance in linking and equating jeopardize the fairness and validity of test scores, and pose particular problems for test-based accountability programs that…
Descriptors: Equated Scores, Tests, Test Bias, Validity
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Lian, Lim Hooi; Yew, Wun Thiam; Meng, Chew Cheng – International Education Studies, 2014
Currently, in order to reform the Malaysian education system, there have been a number of education policy initiatives launched by the Malaysian Ministry of Education (MOE). All these initiatives have encouraged and inculcated teaching and learning for creativity, critical, innovative and higher-order thinking skills rather than conceptual…
Descriptors: Foreign Countries, Educational Policy, Evaluation Methods, Teacher Competencies
Goldhaber, Dan; Chaplin, Duncan – Center for Education Data & Research, 2012
In a provocative and influential paper, Jesse Rothstein (2010) finds that standard value added models (VAMs) suggest implausible future teacher effects on past student achievement, a finding that obviously cannot be viewed as causal. This is the basis of a falsification test (the Rothstein falsification test) that appears to indicate bias in VAM…
Descriptors: School Effectiveness, Teacher Effectiveness, Achievement Gains, Statistical Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Karami, Hossein – Educational Research and Evaluation, 2013
The search for fairness in language testing is distinct from other areas of educational measurement as the object of measurement, that is, language, is part of the identity of the test takers. So, a host of issues enter the scene when one starts to reflect on how to assess people's language abilities. As the quest for fairness in language testing…
Descriptors: Language Skills, Language Tests, Testing, Culture Fair Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Longford, Nicholas T. – Journal of Educational and Behavioral Statistics, 2014
A method for medical screening is adapted to differential item functioning (DIF). Its essential elements are explicit declarations of the level of DIF that is acceptable and of the loss function that quantifies the consequences of the two kinds of inappropriate classification of an item. Instead of a single level and a single function, sets of…
Descriptors: Test Items, Test Bias, Simulation, Hypothesis Testing
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5