Publication Date
| In 2026 | 0 |
| Since 2025 | 3 |
| Since 2022 (last 5 years) | 7 |
| Since 2017 (last 10 years) | 11 |
Descriptor
Source
Author
| Abulela, Mohammed A. A. | 1 |
| B. Barbot | 1 |
| B. Goecke | 1 |
| Bebermeier, Sarah | 1 |
| Chun Wang | 1 |
| Clavel, Jose G. | 1 |
| David Kaplan | 1 |
| David Rutkowski | 1 |
| Davis, Richard L. | 1 |
| Domingue, Benjamin W. | 1 |
| Flannery, Darragh | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 9 |
| Reports - Research | 9 |
| Reports - Evaluative | 2 |
| Speeches/Meeting Papers | 1 |
Education Level
| Secondary Education | 11 |
| Elementary Secondary Education | 1 |
| Higher Education | 1 |
| Postsecondary Education | 1 |
Audience
Location
| Ireland | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| Program for International… | 11 |
| MacArthur Communicative… | 1 |
| Trends in International… | 1 |
What Works Clearinghouse Rating
Mingya Huang; David Kaplan – Journal of Educational and Behavioral Statistics, 2025
The issue of model uncertainty has been gaining interest in education and the social sciences community over the years, and the dominant methods for handling model uncertainty are based on Bayesian inference, particularly, Bayesian model averaging. However, Bayesian model averaging assumes that the true data-generating model is within the…
Descriptors: Bayesian Statistics, Hierarchical Linear Modeling, Statistical Inference, Predictor Variables
Leslie Rutkowski; David Rutkowski – Journal of Creative Behavior, 2025
The Programme for International Student Assessment (PISA) introduced creative thinking as an innovative domain in 2022. This paper examines the unique methodological issues in international assessments and the implications of measuring creative thinking within PISA's framework, including stratified sampling, rotated form designs, and a distinct…
Descriptors: Creativity, Creative Thinking, Measurement, Sampling
B. Goecke; S. Weiss; B. Barbot – Journal of Creative Behavior, 2025
The present paper questions the content validity of the eight creativity-related self-report scales available in PISA 2022's context questionnaire and provides a set of considerations for researchers interested in using these indexes. Specifically, we point out some threats to the content validity of these scales (e.g., "creative thinking…
Descriptors: Creativity, Creativity Tests, Questionnaires, Content Validity
Flannery, Darragh; Gilleece, Lorraine; Clavel, Jose G. – Large-scale Assessments in Education, 2023
Background: The existence of a multiplier, compositional or social context effect is debated extensively in the literature on school effectiveness and also relates to the wider issue of equity in educational outcomes. However, comparatively little attention has been given to whether or not the association between student achievement and school…
Descriptors: Foreign Countries, Secondary School Students, International Assessment, Achievement Tests
Sainan Xu; Jing Lu; Jiwei Zhang; Chun Wang; Gongjun Xu – Grantee Submission, 2024
With the growing attention on large-scale educational testing and assessment, the ability to process substantial volumes of response data becomes crucial. Current estimation methods within item response theory (IRT), despite their high precision, often pose considerable computational burdens with large-scale data, leading to reduced computational…
Descriptors: Educational Assessment, Bayesian Statistics, Statistical Inference, Item Response Theory
Lundgren, Erik – Journal of Educational Data Mining, 2022
Response process data have the potential to provide a rich description of test-takers' thinking processes. However, retrieving insights from these data presents a challenge for educational assessments and educational data mining as they are complex and not well annotated. The present study addresses this challenge by developing a computational…
Descriptors: Problem Solving, Classification, Accuracy, Foreign Countries
Shear, Benjamin R. – Journal of Educational Measurement, 2023
Large-scale standardized tests are regularly used to measure student achievement overall and for student subgroups. These uses assume tests provide comparable measures of outcomes across student subgroups, but prior research suggests score comparisons across gender groups may be complicated by the type of test items used. This paper presents…
Descriptors: Gender Bias, Item Analysis, Test Items, Achievement Tests
Abulela, Mohammed A. A.; Harwell, Michael M. – Educational Sciences: Theory and Practice, 2020
Data analysis is a significant methodological component when conducting quantitative education studies. Guidelines for conducting data analyses in quantitative education studies are common but often underemphasize four important methodological components impacting the validity of inferences: quality of constructed measures, proper handling of…
Descriptors: Educational Research, Educational Researchers, Novices, Data Analysis
Grund, Simon; Lüdtke, Oliver; Robitzsch, Alexander – Journal of Educational and Behavioral Statistics, 2021
Large-scale assessments (LSAs) use Mislevy's "plausible value" (PV) approach to relate student proficiency to noncognitive variables administered in a background questionnaire. This method requires background variables to be completely observed, a requirement that is seldom fulfilled. In this article, we evaluate and compare the…
Descriptors: Data Analysis, Error of Measurement, Research Problems, Statistical Inference
Bebermeier, Sarah; Hagemann, Anne – Teaching of Psychology, 2019
We describe how students can be encouraged to actively review course contents on inferential statistics by creating application-oriented exercises and sample solutions on the basis of concrete and realistic research articles and their data. For evaluation purposes, we use students' reactions to the activity and investigate its effects on the final…
Descriptors: Statistics, Course Content, Statistical Inference, Learning Activities
Wu, Mike; Davis, Richard L.; Domingue, Benjamin W.; Piech, Chris; Goodman, Noah – International Educational Data Mining Society, 2020
Item Response Theory (IRT) is a ubiquitous model for understanding humans based on their responses to questions, used in fields as diverse as education, medicine and psychology. Large modern datasets offer opportunities to capture more nuances in human behavior, potentially improving test scoring and better informing public policy. Yet larger…
Descriptors: Item Response Theory, Accuracy, Data Analysis, Public Policy

Peer reviewed
Direct link
