NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20260
Since 20250
Since 2022 (last 5 years)0
Since 2017 (last 10 years)1
Since 2007 (last 20 years)11
Publication Type
Reports - Evaluative13
Journal Articles12
Audience
What Works Clearinghouse Rating
Showing all 13 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Carpenter, Rachel; Alloway, Tracy – Journal of Psychoeducational Assessment, 2019
School systems across the country are transitioning from paper-based testing (PBT) to computer-based testing (CBT). As this technological shift occurs, more research is necessary to understand the practical and performance implications of administering CBTs. Currently, there is a paucity of research using CBTs to examine working memory (WM)…
Descriptors: Computer Assisted Testing, Test Format, Short Term Memory, Cognitive Processes
Peer reviewed Peer reviewed
Direct linkDirect link
Babiar, Tasha Calvert – Journal of Applied Measurement, 2011
Traditionally, women and minorities have not been fully represented in science and engineering. Numerous studies have attributed these differences to gaps in science achievement as measured by various standardized tests. Rather than describe mean group differences in science achievement across multiple cultures, this study focused on an in-depth…
Descriptors: Test Bias, Science Achievement, Standardized Tests, Grade 8
Kim, Sooyeon; Walker, Michael E. – Educational Testing Service, 2011
This study examines the use of subpopulation invariance indices to evaluate the appropriateness of using a multiple-choice (MC) item anchor in mixed-format tests, which include both MC and constructed-response (CR) items. Linking functions were derived in the nonequivalent groups with anchor test (NEAT) design using an MC-only anchor set for 4…
Descriptors: Test Format, Multiple Choice Tests, Test Items, Gender Differences
Peer reviewed Peer reviewed
Direct linkDirect link
Kalaycioglu, Dilara Bakan; Berberoglu, Giray – Journal of Psychoeducational Assessment, 2011
This study is aimed to detect differential item functioning (DIF) items across gender groups, analyze item content for the possible sources of DIF, and eventually investigate the effect of DIF items on the criterion-related validity of the test scores in the quantitative section of the university entrance examination (UEE) in Turkey. The reason…
Descriptors: Test Bias, College Entrance Examinations, Item Analysis, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Pae, Tae-Il – Language Testing, 2012
This study tracked gender differential item functioning (DIF) on the English subtest of the Korean College Scholastic Aptitude Test (KCSAT) over a nine-year period across three data points, using both the Mantel-Haenszel (MH) and item response theory likelihood ratio (IRT-LR) procedures. Further, the study identified two factors (i.e. reading…
Descriptors: Aptitude Tests, Academic Aptitude, Language Tests, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Ou Lydia; Wilson, Mark – Applied Measurement in Education, 2009
Many efforts have been made to determine and explain differential gender performance on large-scale mathematics assessments. A well-agreed-on conclusion is that gender differences are contextualized and vary across math domains. This study investigated the pattern of gender differences by item domain (e.g., Space and Shape, Quantity) and item type…
Descriptors: Gender Differences, Mathematics Tests, Measurement, Test Format
Peer reviewed Peer reviewed
Direct linkDirect link
Lohman, David F.; Lakin, Joni M. – British Journal of Educational Psychology, 2009
Background: Strand, Deary, and Smith (2006) reported an analysis of sex differences on the Cognitive Abilities Test (CAT) for over 320,000 UK students 11-12 years old. Although mean differences were small, males were overrepresented at the upper and lower extremes of the score distributions on the quantitative and non-verbal batteries and at the…
Descriptors: Gender Differences, Cognitive Tests, Foreign Countries, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Lafontaine, Dominique; Monseur, Christian – European Educational Research Journal, 2009
In this article we discuss how apparently indicators that may appear straightforward, such as gender differences, need to be interpreted with extreme care. In particular, we consider how the assessment framework, and the methodology of international surveys, may have a potential impact on the results and on the indicators. Through analysis of…
Descriptors: Foreign Countries, Reading Comprehension, Test Format, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Einarsdottir, Sif; Rounds, James – Journal of Vocational Behavior, 2009
Item response theory was used to address gender bias in interest measurement. Differential item functioning (DIF) technique, SIBTEST and DIMTEST for dimensionality, were applied to the items of the six General Occupational Theme (GOT) and 25 Basic Interest (BI) scales in the Strong Interest Inventory. A sample of 1860 women and 1105 men was used.…
Descriptors: Test Format, Females, Vocational Interests, Construct Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Dorans, Neil J.; Liu, Jinghua; Hammond, Shelby – Applied Psychological Measurement, 2008
This exploratory study was built on research spanning three decades. Petersen, Marco, and Stewart (1982) conducted a major empirical investigation of the efficacy of different equating methods. The studies reported in Dorans (1990) examined how different equating methods performed across samples selected in different ways. Recent population…
Descriptors: Test Format, Equated Scores, Sampling, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Prestera, Gustavo E.; Clariana, Roy; Peck, Andrew – Journal of Educational Multimedia and Hypermedia, 2005
In this experimental study, 44 undergraduates completed five computer-based instructional lessons and either two multiplechoice tests or two fill-in-the-blank tests. Color-coded borders were displayed during the lesson, adjacent to the screen text and illustrations. In the experimental condition, corresponding border colors were shown at posttest.…
Descriptors: Experimental Groups, Computer Assisted Instruction, Instructional Effectiveness, Multiple Choice Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Wiliam, Dylan – Review of Research in Education, 2010
The idea that validity should be considered a property of inferences, rather than of assessments, has developed slowly over the past century. In early writings about the validity of educational assessments, validity was defined as a property of an assessment. The most common definition was that an assessment was valid to the extent that it…
Descriptors: Educational Assessment, Validity, Inferences, Construct Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Pae, Tae-Il – System: An International Journal of Educational Technology and Applied Linguistics, 2004
This paper examines the effect of gender on English reading comprehension for Korean EFL (English as a Foreign Language) learners. The gender effect was measured using a DIF (Differential Item Functioning) methodology. Specifically, gender DIF was investigated for a random sample of 14,000 Korean examinees (7,000 males and 7,000 females) who took…
Descriptors: Reading Comprehension, Test Format, Content Analysis, Gender Differences