Publication Date
| In 2026 | 0 |
| Since 2025 | 1 |
| Since 2022 (last 5 years) | 1 |
| Since 2017 (last 10 years) | 2 |
| Since 2007 (last 20 years) | 3 |
Descriptor
| Computer Assisted Testing | 3 |
| Prediction | 3 |
| Scoring | 2 |
| Algorithms | 1 |
| Automation | 1 |
| Bias | 1 |
| Comparative Analysis | 1 |
| Correlation | 1 |
| Difficulty Level | 1 |
| Elementary School Students | 1 |
| Essays | 1 |
| More ▼ | |
Source
| Journal of Educational… | 3 |
Author
| Albano, Anthony D. | 1 |
| Cai, Liuhan | 1 |
| Ikkyu Choi | 1 |
| Lease, Erin M. | 1 |
| Lottridge, Sue | 1 |
| Matthew S. Johnson | 1 |
| Mayfield, Elijah | 1 |
| McConnell, Scott R. | 1 |
| Shermis, Mark D. | 1 |
Publication Type
| Journal Articles | 3 |
| Reports - Research | 3 |
Education Level
| Elementary Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Ikkyu Choi; Matthew S. Johnson – Journal of Educational Measurement, 2025
Automated scoring systems provide multiple benefits but also pose challenges, notably potential bias. Various methods exist to evaluate these algorithms and their outputs for bias. Upon detecting bias, the next logical step is to investigate its cause, often by examining feature distributions. Recently, Johnson and McCaffrey proposed an…
Descriptors: Prediction, Bias, Automation, Scoring
Albano, Anthony D.; Cai, Liuhan; Lease, Erin M.; McConnell, Scott R. – Journal of Educational Measurement, 2019
Studies have shown that item difficulty can vary significantly based on the context of an item within a test form. In particular, item position may be associated with practice and fatigue effects that influence item parameter estimation. The purpose of this research was to examine the relevance of item position specifically for assessments used in…
Descriptors: Test Items, Computer Assisted Testing, Item Analysis, Difficulty Level
Shermis, Mark D.; Lottridge, Sue; Mayfield, Elijah – Journal of Educational Measurement, 2015
This study investigated the impact of anonymizing text on predicted scores made by two kinds of automated scoring engines: one that incorporates elements of natural language processing (NLP) and one that does not. Eight data sets (N = 22,029) were used to form both training and test sets in which the scoring engines had access to both text and…
Descriptors: Scoring, Essays, Computer Assisted Testing, Natural Language Processing

Peer reviewed
Direct link
