NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 26 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Silvia Testa; Renato Miceli; Renato Miceli – Educational Measurement: Issues and Practice, 2025
Random Equating (RE) and Heuristic Approach (HA) are two linking procedures that may be used to compare the scores of individuals in two tests that measure the same latent trait, in conditions where there are no common items or individuals. In this study, RE--that may only be used when the individuals taking the two tests come from the same…
Descriptors: Comparative Testing, Heuristics, Problem Solving, Personality Traits
Peer reviewed Peer reviewed
Direct linkDirect link
Jung Yeon Park; Sean Joo; Zikun Li; Hyejin Yoon – Educational Measurement: Issues and Practice, 2025
This study examines potential assessment bias based on students' primary language status in PISA 2018. Specifically, multilingual (MLs) and nonmultilingual (non-MLs) students in the United States are compared with regard to their response time as well as scored responses across three cognitive domains (reading, mathematics, and science).…
Descriptors: Achievement Tests, Secondary School Students, International Assessment, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Xiao, Yue; Veldkamp, Bernard; Liu, Hongyun – Educational Measurement: Issues and Practice, 2022
The action sequences of respondents in problem-solving tasks reflect rich and detailed information about their performance, including differences in problem-solving ability, even if item scores are equal. It is therefore not sufficient to infer individual problem-solving skills based solely on item scores. This study is a preliminary attempt to…
Descriptors: Problem Solving, Item Response Theory, Scores, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Sooyeon; Walker, Michael E. – Educational Measurement: Issues and Practice, 2022
Test equating requires collecting data to link the scores from different forms of a test. Problems arise when equating samples are not equivalent and the test forms to be linked share no common items by which to measure or adjust for the group nonequivalence. Using data from five operational test forms, we created five pairs of research forms for…
Descriptors: Ability, Tests, Equated Scores, Testing Problems
Peer reviewed Peer reviewed
Direct linkDirect link
Skaggs, Gary; Hein, Serge F.; Wilkins, Jesse L. M. – Educational Measurement: Issues and Practice, 2020
In test-centered standard-setting methods, borderline performance can be represented by many different profiles of strengths and weaknesses. As a result, asking panelists to estimate item or test performance for a hypothetical group study of borderline examinees, or a typical borderline examinee, may be an extremely difficult task and one that can…
Descriptors: Standard Setting (Scoring), Cutting Scores, Testing Problems, Profiles
Peer reviewed Peer reviewed
Direct linkDirect link
Wyse, Adam E.; Babcock, Ben – Educational Measurement: Issues and Practice, 2020
A common belief is that the Bookmark method is a cognitively simpler standard-setting method than the modified Angoff method. However, a limited amount of research has investigated panelist's ability to perform well the Bookmark method, and whether some of the challenges panelists face with the Angoff method may also be present in the Bookmark…
Descriptors: Standard Setting (Scoring), Evaluation Methods, Testing Problems, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Moon, Jung Aa; Keehner, Madeleine; Katz, Irvin R. – Educational Measurement: Issues and Practice, 2019
The current study investigated how item formats and their inherent affordances influence test-takers' cognition under uncertainty. Adult participants solved content-equivalent math items in multiple-selection multiple-choice and four alternative grid formats. The results indicated that participants' affirmative response tendency (i.e., judge the…
Descriptors: Affordances, Test Items, Test Format, Test Wiseness
Peer reviewed Peer reviewed
Direct linkDirect link
Wind, Stefanie A.; Schumacker, Randall E. – Educational Measurement: Issues and Practice, 2017
The term measurement disturbance has been used to describe systematic conditions that affect a measurement process, resulting in a compromised interpretation of person or item estimates. Measurement disturbances have been discussed in relation to systematic response patterns associated with items and persons, such as start-up, plodding, boredom,…
Descriptors: Measurement, Testing Problems, Writing Tests, Performance Based Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
An, Chen; Braun, Henry; Walsh, Mary E. – Educational Measurement: Issues and Practice, 2018
Making causal inferences from a quasi-experiment is difficult. Sensitivity analysis approaches to address hidden selection bias thus have gained popularity. This study serves as an introduction to a simple but practical form of sensitivity analysis using Monte Carlo simulation procedures. We examine estimated treatment effects for a school-based…
Descriptors: Statistical Inference, Intervention, Program Effectiveness, Quasiexperimental Design
Peer reviewed Peer reviewed
Direct linkDirect link
Penfield, Randall D.; Gattamorta, Karina; Childs, Ruth A. – Educational Measurement: Issues and Practice, 2009
Traditional methods for examining differential item functioning (DIF) in polytomously scored test items yield a single item-level index of DIF and thus provide no information concerning which score levels are implicated in the DIF effect. To address this limitation of DIF methodology, the framework of differential step functioning (DSF) has…
Descriptors: Test Bias, Test Items, Evaluation Methods, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Ferrara, Steve; Svetina, Dubravka; Skucha, Sylvia; Davidson, Anne H. – Educational Measurement: Issues and Practice, 2011
Items on test score scales located at and below the Proficient cut score define the content area knowledge and skills required to achieve proficiency. Alternately, examinees who perform at the Proficient level on a test can be expected to be able to demonstrate that they have mastered most of the knowledge and skills represented by the items at…
Descriptors: Knowledge Level, Mathematics Tests, Program Effectiveness, Inferences
Peer reviewed Peer reviewed
Direct linkDirect link
Kingston, Neal; Nash, Brooke – Educational Measurement: Issues and Practice, 2011
An effect size of about 0.70 (or 0.40-0.70) is often claimed for the efficacy of formative assessment, but is not supported by the existing research base. More than 300 studies that appeared to address the efficacy of formative assessment in grades K-12 were reviewed. Many of the studies had severely flawed research designs yielding…
Descriptors: Elementary Secondary Education, Formative Evaluation, Program Effectiveness, Effect Size
Peer reviewed Peer reviewed
Moran, Mary Ross; And Others – Educational Measurement: Issues and Practice, 1991
Practices identified by experts as critical variables in eliciting writing samples were checked against 12 randomly selected studies using holistic ratings to derive descriptions of inferential statistical results for described samples. The studies often lacked precise information about these variables, limiting understanding of writing evaluation…
Descriptors: Cues, Educational Practices, Examiners, Holistic Evaluation
Peer reviewed Peer reviewed
Phillips, S. E.; Clarizio, Harvey F. – Educational Measurement: Issues and Practice, 1988
Two major problems related to the identification of learning disabilities with individually administered achievement tests are discussed: (1) the appropriateness of standard versus developmental scores for determining the severity of discrepancy; and (2) the limitations of existing developmental score scales. Characteristics of the developmental…
Descriptors: Achievement Tests, Diagnostic Tests, Learning Disabilities, Scores
Peer reviewed Peer reviewed
Bauer, Ernest A. – Educational Measurement: Issues and Practice, 1985
Misreadings of pencil answer marks on test answer sheets by optical scanners cause scoring errors. Twenty-six different pencils were tested for readability differences when optically scanned. Complete light and dark marks scanned perfectly for 18 pencils. Totals for all six mark types ranged from 947 to 1726 out of 1800. (BS)
Descriptors: Answer Sheets, Elementary Secondary Education, Error of Measurement, Optical Scanners
Previous Page | Next Page ยป
Pages: 1  |  2