Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 3 |
| Since 2017 (last 10 years) | 10 |
| Since 2007 (last 20 years) | 21 |
Descriptor
| Achievement Tests | 43 |
| Language Tests | 43 |
| Test Items | 43 |
| Foreign Countries | 18 |
| Language Proficiency | 18 |
| Language Skills | 16 |
| English (Second Language) | 14 |
| Second Language Learning | 14 |
| French | 13 |
| Test Construction | 13 |
| Academic Achievement | 12 |
| More ▼ | |
Source
Author
| Abbott, Marilyn L. | 1 |
| Abedi, Jamal | 1 |
| Ali, Usama | 1 |
| Anderson, Michael | 1 |
| Bachman, Patina L. | 1 |
| Baghaei, Purya | 1 |
| Brown, Terran | 1 |
| Bynum, Bethany H. | 1 |
| Cai, Li | 1 |
| Cemile Dogan | 1 |
| Chang, Sandy M. | 1 |
| More ▼ | |
Publication Type
Education Level
Audience
| Practitioners | 7 |
| Teachers | 5 |
| Administrators | 2 |
| Students | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| Measures of Academic Progress | 2 |
| College Board Achievement… | 1 |
| Iowa Tests of Basic Skills | 1 |
What Works Clearinghouse Rating
Baghaei, Purya; Christensen, Karl Bang – Language Testing, 2023
C-tests are gap-filling tests mainly used as rough and economical measures of second-language proficiency for placement and research purposes. A C-test usually consists of several short independent passages where the second half of every other word is deleted. Owing to their interdependent structure, C-test items violate the local independence…
Descriptors: Item Response Theory, Language Tests, Language Proficiency, Second Language Learning
Cemile Dogan – International Journal of Contemporary Educational Research, 2023
Tests are designed as an integral part of the teaching process, necessarily including stakeholders from the onset of preparations to grade allocation, the administration of the test, and the interpretation of the results. The process commences with selecting content to evaluate, deciding upon the skills to be tested, and to meet course objectives…
Descriptors: Teacher Made Tests, Language Tests, Second Language Instruction, English (Second Language)
Chung, Seungwon; Cai, Li – Journal of Educational and Behavioral Statistics, 2021
In the research reported here, we propose a new method for scale alignment and test scoring in the context of supporting students with disabilities. In educational assessment, students from these special populations take modified tests because of a demonstrated disability that requires more assistance than standard testing accommodation. Updated…
Descriptors: Students with Disabilities, Scoring, Achievement Tests, Test Items
Russell, Michael; Szendey, Olivia; Kaplan, Larry – Educational Assessment, 2021
Differential Item Function (DIF) analysis is commonly employed to examine potential bias produced by a test item. Since its introduction DIF analyses have focused on potential bias related to broad categories of oppression, including gender, racial stratification, economic class, and ableness. More recently, efforts to examine the effects of…
Descriptors: Test Bias, Achievement Tests, Individual Characteristics, Disadvantaged
Shafipoor, Mahdieh; Ravand, Hamdollah; Maftoon, Parviz – Language Testing in Asia, 2021
The current study compared the model fit indices, skill mastery probabilities, and classification accuracy of six Diagnostic Classification Models (DCMs): a general model (G-DINA) against five specific models (LLM, RRUM, ACDM, DINA, and DINO). To do so, the response data to the grammar and vocabulary sections of a General English Achievement Test,…
Descriptors: Goodness of Fit, Models, Classification, Grammar
He, Wei – NWEA, 2022
To ensure that student academic growth in a subject area is accurately captured, it is imperative that the underlying scale remains stable over time. As item parameter stability constitutes one of the factors that affects scale stability, NWEA® periodically conducts studies to check for the stability of the item parameter estimates for MAP®…
Descriptors: Achievement Tests, Test Items, Test Reliability, Academic Achievement
Jamalzadeh, Mehri; Lotfi, Ahmad Reza; Rostami, Masoud – Language Testing in Asia, 2021
The current study sought to examine the validity of a General English Achievement Test (GEAT), administered to university students in the fall semester of 2018-2019 academic year, by hybridizing differential information (DIF) and differential distractor function (DDF) analytical models. Using a purposive sampling method, from the target population…
Descriptors: Language Tests, Achievement Tests, Undergraduate Students, Islam
Steedle, Jeffrey T.; Morrison, Kristin M. – Educational Assessment, 2019
Assessment items are commonly field tested prior to operational use to observe statistical item properties such as difficulty. Item parameter estimates from field testing may be used to assign scores via pre-equating or computer adaptive designs. This study examined differences between item difficulty estimates based on field test and operational…
Descriptors: Field Tests, Test Items, Statistics, Difficulty Level
Löwenadler, John – Language Testing, 2019
This study aims to investigate patterns of variation in the interplay of L2 language ability and general reading comprehension skills in L2 reading, by comparing item-level effects of test-takers' results on L1 and L2 reading comprehension tests. The material comes from more than 500,000 people tested on L1 (Swedish) and L2 (English) in the…
Descriptors: Swedish, English (Second Language), Second Language Learning, Second Language Instruction
Li, Sylvia; Meyer, Patrick – NWEA, 2019
This simulation study examines the measurement precision, item exposure rates, and the depth of the MAP® Growth™ item pools under various grade-level restrictions. Unlike most summative assessments, MAP Growth allows examinees to see items from any grade level, regardless of the examinee's actual grade level. It does not limit the test to items…
Descriptors: Achievement Tests, Item Banks, Test Items, Instructional Program Divisions
Ferrara, Steve; Steedle, Jeffrey; Kinsman, Amy – Partnership for Assessment of Readiness for College and Careers, 2015
We report results from the following three analyses of PARCC [Partnership for Assessment of Readiness for College and Careers] cognitive complexity measures, based on 2014 field test item and task development and field test data. We conducted classification and regression tree analyses using 2014 PARCC field test data to do the following: (1)…
Descriptors: Cognitive Processes, Difficulty Level, Test Items, Mathematics Tests
Liu, Junhui; Brown, Terran; Chen, Jianshen; Ali, Usama; Hou, Likun; Costanzo, Kate – Partnership for Assessment of Readiness for College and Careers, 2016
The Partnership for Assessment of Readiness for College and Careers (PARCC) is a state-led consortium working to develop next-generation assessments that more accurately, compared to previous assessments, measure student progress toward college and career readiness. The PARCC assessments include both English Language Arts/Literacy (ELA/L) and…
Descriptors: Testing, Achievement Tests, Test Items, Test Bias
Thacker, Arthur A.; Dickinson, Emily R.; Bynum, Bethany H.; Wen, Yao; Smith, Erin; Sinclair, Andrea L.; Deatz, Richard C.; Wise, Lauress L. – Partnership for Assessment of Readiness for College and Careers, 2015
The Partnership for Assessment of Readiness for College and Careers (PARCC) field tests during the spring of 2014 provided an opportunity to investigate the quality of the items, tasks, and associated stimuli. HumRRO conducted several research studies summarized in this report. Quality of test items is integral to the "Theory of Action"…
Descriptors: Achievement Tests, Test Items, Common Core State Standards, Difficulty Level
Holster, Trevor A.; Lake, J. – Language Assessment Quarterly, 2016
Stewart questioned Beglar's use of Rasch analysis of the Vocabulary Size Test (VST) and advocated the use of 3-parameter logistic item response theory (3PLIRT) on the basis that it models a non-zero lower asymptote for items, often called a "guessing" parameter. In support of this theory, Stewart presented fit statistics derived from…
Descriptors: Guessing (Tests), Item Response Theory, Vocabulary, Language Tests
Finch, W. Holmes; Hernández Finch, Maria E. – Educational and Psychological Measurement, 2013
The assessment of test data for the presence of differential item functioning (DIF) is a key component of instrument development and validation. Among the many methods that have been used successfully in such analyses is the mixture modeling approach. Using this approach to identify the presence of DIF has been touted as potentially superior for…
Descriptors: Learning Disabilities, Testing Accommodations, Test Bias, Item Response Theory

Peer reviewed
Direct link
