Publication Date
| In 2026 | 0 |
| Since 2025 | 11 |
| Since 2022 (last 5 years) | 44 |
Descriptor
| Test Items | 44 |
| Data Analysis | 23 |
| Item Response Theory | 17 |
| Data Collection | 13 |
| Models | 11 |
| Test Construction | 11 |
| Scores | 9 |
| Artificial Intelligence | 8 |
| Item Analysis | 8 |
| Accuracy | 7 |
| Data Use | 7 |
| More ▼ | |
Source
Author
| Amanda Goodwin | 2 |
| Chun Wang | 2 |
| Gongjun Xu | 2 |
| K. Kawena Begay | 2 |
| Matthew Naveiras | 2 |
| Miranda Kucera | 2 |
| Paul De Boeck | 2 |
| Sun-Joo Cho | 2 |
| Allan S. Cohen | 1 |
| Ames, Allison J. | 1 |
| An, Lily Shiao | 1 |
| More ▼ | |
Publication Type
Education Level
| Secondary Education | 7 |
| Higher Education | 5 |
| Postsecondary Education | 5 |
| Junior High Schools | 4 |
| Middle Schools | 4 |
| Elementary Education | 2 |
| High Schools | 2 |
| Elementary Secondary Education | 1 |
| Grade 11 | 1 |
| Grade 8 | 1 |
Audience
| Administrators | 1 |
| Policymakers | 1 |
| Researchers | 1 |
| Teachers | 1 |
Location
| Florida | 1 |
| Germany | 1 |
| New Zealand | 1 |
| Oklahoma (Tulsa) | 1 |
| Thailand | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| ACTFL Oral Proficiency… | 1 |
| Big Five Inventory | 1 |
| California Critical Thinking… | 1 |
| Cornell Critical Thinking Test | 1 |
| Hopkins Symptom Checklist | 1 |
| Trends in International… | 1 |
| Watson Glaser Critical… | 1 |
What Works Clearinghouse Rating
Stella Y. Kim; Sungyeun Kim – Educational Measurement: Issues and Practice, 2025
This study presents several multivariate Generalizability theory designs for analyzing automatic item-generated (AIG) based test forms. The study used real data to illustrate the analysis procedure and discuss practical considerations. We collected the data from two groups of students, each group receiving a different form generated by AIG. A…
Descriptors: Generalizability Theory, Automation, Test Items, Students
Miranda Kucera; K. Kawena Begay – Communique, 2025
While the field advocates for a diversified and comprehensive professional role (National Association of School Psychologists, 2020), school psychologists have long spent most of their time in assessment-related activities (Farmer et al., 2021), averaging about eight cognitive evaluations monthly (Benson et al., 2020). Assessment practices have…
Descriptors: Equal Education, Student Evaluation, Evaluation Methods, Standardized Tests
Miranda Kucera; K. Kawena Begay – Communique, 2025
In Part 1 of this series, the authors briefly reviewed some challenges inherent in using standardized tools with students who are not well represented in norming data. To help readers clearly conceptualize the framework steps, the authors present two case studies that showcase how a nonstandardized approach to assessment can be individualized to…
Descriptors: Equal Education, Student Evaluation, Evaluation Methods, Standardized Tests
Marcoulides, Katerina M. – Measurement: Interdisciplinary Research and Perspectives, 2023
Integrative data analyses have recently been shown to be an effective tool for researchers interested in synthesizing datasets from multiple studies in order to draw statistical or substantive conclusions. The actual process of integrating the different datasets depends on the availability of some common measures or items reflecting the same…
Descriptors: Data Analysis, Synthesis, Test Items, Simulation
Engelhard, George – Educational and Psychological Measurement, 2023
The purpose of this study is to introduce a functional approach for modeling unfolding response data. Functional data analysis (FDA) has been used for examining cumulative item response data, but a functional approach has not been systematically used with unfolding response processes. A brief overview of FDA is presented and illustrated within the…
Descriptors: Data Analysis, Models, Responses, Test Items
Leventhal, Brian C.; Gregg, Nikole; Ames, Allison J. – Measurement: Interdisciplinary Research and Perspectives, 2022
Response styles introduce construct-irrelevant variance as a result of respondents systematically responding to Likert-type items regardless of content. Methods to account for response styles through data analysis as well as approaches to mitigating the effects of response styles during data collection have been well-documented. Recent approaches…
Descriptors: Response Style (Tests), Item Response Theory, Test Items, Likert Scales
Neuert, Cornelia E.; Meitinger, Katharina; Behr, Dorothée – Sociological Methods & Research, 2023
The method of web probing integrates cognitive interviewing techniques into web surveys and is increasingly used to evaluate survey questions. In a usual web probing scenario, probes are administered immediately after the question to be tested (concurrent probing), typically as open-ended questions. A second possibility of administering probes is…
Descriptors: Internet, Online Surveys, Test Items, Evaluation
He, Dan – ProQuest LLC, 2023
This dissertation examines the effectiveness of machine learning algorithms and feature engineering techniques for analyzing process data and predicting test performance. The study compares three classification approaches and identifies item-specific process features that are highly predictive of student performance. The findings suggest that…
Descriptors: Artificial Intelligence, Data Analysis, Algorithms, Classification
Chengcheng Li – ProQuest LLC, 2022
Categorical data become increasingly ubiquitous in the modern big data era. In this dissertation, we propose novel statistical learning and inference methods for large-scale categorical data, focusing on latent variable models and their applications to psychometrics. In psychometric assessments, the subjects' underlying aptitude often cannot be…
Descriptors: Statistical Inference, Data Analysis, Psychometrics, Raw Scores
Jiawei Xiong; George Engelhard; Allan S. Cohen – Measurement: Interdisciplinary Research and Perspectives, 2025
It is common to find mixed-format data results from the use of both multiple-choice (MC) and constructed-response (CR) questions on assessments. Dealing with these mixed response types involves understanding what the assessment is measuring, and the use of suitable measurement models to estimate latent abilities. Past research in educational…
Descriptors: Responses, Test Items, Test Format, Grade 8
Mihyun Son; Minsu Ha – Education and Information Technologies, 2025
Digital literacy is essential for scientific literacy in a digital world. Although the NGSS Practices include many activities that require digital literacy, most studies have examined digital literacy from a generic perspective rather than a curricular context. This study aimed to develop a self-report tool to measure elements of digital literacy…
Descriptors: Test Construction, Measures (Individuals), Digital Literacy, Scientific Literacy
Mingfeng Xue; Ping Chen – Journal of Educational Measurement, 2025
Response styles pose great threats to psychological measurements. This research compares IRTree models and anchoring vignettes in addressing response styles and estimating the target traits. It also explores the potential of combining them at the item level and total-score level (ratios of extreme and middle responses to vignettes). Four models…
Descriptors: Item Response Theory, Models, Comparative Analysis, Vignettes
Wu, Tong; Kim, Stella Y.; Westine, Carl – Educational and Psychological Measurement, 2023
For large-scale assessments, data are often collected with missing responses. Despite the wide use of item response theory (IRT) in many testing programs, however, the existing literature offers little insight into the effectiveness of various approaches to handling missing responses in the context of scale linking. Scale linking is commonly used…
Descriptors: Data Analysis, Responses, Statistical Analysis, Measurement
Hongwen Guo; Matthew S. Johnson; Daniel F. McCaffrey; Lixong Gu – ETS Research Report Series, 2024
The multistage testing (MST) design has been gaining attention and popularity in educational assessments. For testing programs that have small test-taker samples, it is challenging to calibrate new items to replenish the item pool. In the current research, we used the item pools from an operational MST program to illustrate how research studies…
Descriptors: Test Items, Test Construction, Sample Size, Scaling
Andrew D. Ho – Journal of Educational and Behavioral Statistics, 2024
I review opportunities and threats that widely accessible Artificial Intelligence (AI)-powered services present for educational statistics and measurement. Algorithmic and computational advances continue to improve approaches to item generation, scale maintenance, test security, test scoring, and score reporting. Predictable misuses of AI for…
Descriptors: Artificial Intelligence, Measurement, Educational Assessment, Technology Uses in Education

Peer reviewed
Direct link
