Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 1 |
| Since 2017 (last 10 years) | 7 |
| Since 2007 (last 20 years) | 11 |
Descriptor
| Test Format | 11 |
| Test Items | 9 |
| Mathematics Tests | 4 |
| Multiple Choice Tests | 4 |
| Scores | 4 |
| College Students | 3 |
| Foreign Countries | 3 |
| Language Tests | 3 |
| Test Wiseness | 3 |
| Difficulty Level | 2 |
| Elementary School Students | 2 |
| More ▼ | |
Source
| Educational Assessment | 11 |
Author
Publication Type
| Journal Articles | 11 |
| Reports - Research | 10 |
| Information Analyses | 1 |
| Reports - Evaluative | 1 |
Education Level
| Higher Education | 3 |
| Elementary Education | 2 |
| Grade 3 | 2 |
| Intermediate Grades | 2 |
| Middle Schools | 2 |
| Postsecondary Education | 2 |
| Early Childhood Education | 1 |
| Elementary Secondary Education | 1 |
| Grade 4 | 1 |
| Grade 5 | 1 |
| Grade 6 | 1 |
| More ▼ | |
Audience
Laws, Policies, & Programs
Assessments and Surveys
| National Assessment of… | 1 |
What Works Clearinghouse Rating
Wind, Stefanie A.; Guo, Wenjing – Educational Assessment, 2021
Scoring procedures for the constructed-response (CR) items in large-scale mixed-format educational assessments often involve checks for rater agreement or rater reliability. Although these analyses are important, researchers have documented rater effects that persist despite rater training and that are not always detected in rater agreement and…
Descriptors: Scoring, Responses, Test Items, Test Format
Bulut, Okan; Bulut, Hatice Cigdem; Cormier, Damien C.; Ilgun Dibek, Munevver; Sahin Kursad, Merve – Educational Assessment, 2023
Some statewide testing programs allow students to receive corrective feedback and revise their answers during testing. Despite its pedagogical benefits, the effects of providing revision opportunities remain unknown in the context of alternate assessments. Therefore, this study examined student data from a large-scale alternate assessment that…
Descriptors: Error Correction, Alternative Assessment, Feedback (Response), Multiple Choice Tests
Moon, Jung Aa; Keehner, Madeleine; Katz, Irvin R. – Educational Assessment, 2020
We investigated how item formats influence test takers' response tendencies under uncertainty. Adult participants solved content-equivalent math items in three formats: multiple-selection multiple-choice, grid with forced-choice (true-false) options, and grid with non-forced-choice options. Participants showed a greater tendency to commit (rather…
Descriptors: College Students, Test Wiseness, Test Format, Test Items
Becker, Anthony; Nekrasova-Beker, Tatiana – Educational Assessment, 2018
While previous research has identified numerous factors that contribute to item difficulty, studies involving large-scale reading tests have provided mixed results. This study examined five selected-response item types used to measure reading comprehension in the Pearson Test of English Academic: a) multiple-choice (choose one answer), b)…
Descriptors: Reading Comprehension, Test Items, Reading Tests, Test Format
Steedle, Jeffrey T.; Morrison, Kristin M. – Educational Assessment, 2019
Assessment items are commonly field tested prior to operational use to observe statistical item properties such as difficulty. Item parameter estimates from field testing may be used to assign scores via pre-equating or computer adaptive designs. This study examined differences between item difficulty estimates based on field test and operational…
Descriptors: Field Tests, Test Items, Statistics, Difficulty Level
Hassler Hallstedt, Martin; Ghaderi, Ata – Educational Assessment, 2018
Tablets can be used to facilitate systematic testing of academic skills. Yet, when using validated paper tests on tablet, comparability between the mediums must be established. Comparability between a tablet and a paper version of a basic math skills test (HRT: Heidelberger Rechen Test 1-4) was investigated. Five samples with second and third…
Descriptors: Handheld Devices, Scores, Test Format, Computer Assisted Testing
Kan, Adnan; Bulut, Okan; Cormier, Damien C. – Educational Assessment, 2019
Item stem formats can alter the cognitive complexity as well as the type of abilities required for solving mathematics items. Consequently, it is possible that item stem formats can affect the dimensional structure of mathematics assessments. This empirical study investigated the relationship between item stem format and the dimensionality of…
Descriptors: Mathematics Tests, Test Items, Test Format, Problem Solving
Lakin, Joni M. – Educational Assessment, 2014
The purpose of test directions is to familiarize examinees with a test so that they respond to items in the manner intended. However, changes in educational measurement as well as the U.S. student population present new challenges to test directions and increase the impact that differential familiarity could have on the validity of test score…
Descriptors: Test Content, Test Construction, Best Practices, Familiarity
Pae, Hye K. – Educational Assessment, 2014
This study investigated the role of item formats in the performance of 206 nonnative speakers of English on expressive skills (i.e., speaking and writing). Test scores were drawn from the field test of the "Pearson Test of English Academic" for Chinese, French, Hebrew, and Korean native speakers. Four item formats, including…
Descriptors: Test Items, Test Format, Speech Skills, Writing Skills
Cheng, Liying; DeLuca, Christopher – Educational Assessment, 2011
Test-takers' interpretations of validity as related to test constructs and test use have been widely debated in large-scale language assessment. This study contributes further evidence to this debate by examining 59 test-takers' perspectives in writing large-scale English language tests. Participants wrote about their test-taking experiences in…
Descriptors: Language Tests, Test Validity, Test Use, English
DeMars, Christine E. – Educational Assessment, 2007
A series of 8 tests was administered to university students over 4 weeks for program assessment purposes. The stakes of these tests were low for students; they received course points based on test completion, not test performance. Tests were administered in a counterbalanced order across 2 administrations. Response time effort, a measure of the…
Descriptors: Reaction Time, Guessing (Tests), Testing Programs, College Students

Peer reviewed
Direct link
