Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 3 |
Descriptor
Test Format | 3 |
Medical Students | 2 |
Responses | 2 |
Test Construction | 2 |
Test Items | 2 |
Artificial Intelligence | 1 |
Automation | 1 |
College Graduates | 1 |
Computer Assisted Testing | 1 |
Computer Uses in Education | 1 |
Correlation | 1 |
More ▼ |
Source
Applied Measurement in… | 3 |
Author
Boulais, André-Philippe | 1 |
Brian E. Clauser | 1 |
Carney, Michele | 1 |
Cavey, Laurie | 1 |
De Champlain, André | 1 |
Gierl, Mark J. | 1 |
Janet Mee | 1 |
Lai, Hollis | 1 |
Le An Ha | 1 |
Mo, Ya | 1 |
Peter Baldwin | 1 |
More ▼ |
Publication Type
Journal Articles | 3 |
Reports - Research | 3 |
Education Level
Higher Education | 3 |
Postsecondary Education | 3 |
Audience
Location
Canada | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Brian E. Clauser; Victoria Yaneva; Peter Baldwin; Le An Ha; Janet Mee – Applied Measurement in Education, 2024
Multiple-choice questions have become ubiquitous in educational measurement because the format allows for efficient and accurate scoring. Nonetheless, there remains continued interest in constructed-response formats. This interest has driven efforts to develop computer-based scoring procedures that can accurately and efficiently score these items.…
Descriptors: Computer Uses in Education, Artificial Intelligence, Scoring, Responses
Mo, Ya; Carney, Michele; Cavey, Laurie; Totorica, Tatia – Applied Measurement in Education, 2021
There is a need for assessment items that assess complex constructs but can also be efficiently scored for evaluation of teacher education programs. In an effort to measure the construct of teacher attentiveness in an efficient and scalable manner, we are using exemplar responses elicited by constructed-response item prompts to develop…
Descriptors: Protocol Analysis, Test Items, Responses, Mathematics Teachers
Gierl, Mark J.; Lai, Hollis; Pugh, Debra; Touchie, Claire; Boulais, André-Philippe; De Champlain, André – Applied Measurement in Education, 2016
Item development is a time- and resource-intensive process. Automatic item generation integrates cognitive modeling with computer technology to systematically generate test items. To date, however, items generated using cognitive modeling procedures have received limited use in operational testing situations. As a result, the psychometric…
Descriptors: Psychometrics, Multiple Choice Tests, Test Items, Item Analysis