Publication Date
| In 2026 | 0 |
| Since 2025 | 23 |
| Since 2022 (last 5 years) | 134 |
Descriptor
| Simulation | 134 |
| Item Response Theory | 67 |
| Feedback (Response) | 36 |
| Models | 36 |
| Test Items | 33 |
| Foreign Countries | 21 |
| Item Analysis | 21 |
| Accuracy | 20 |
| Algorithms | 20 |
| Responses | 18 |
| Error of Measurement | 17 |
| More ▼ | |
Source
Author
| Chun Wang | 5 |
| Amanda Goodwin | 4 |
| Jason M. Harley | 4 |
| Matthew Naveiras | 4 |
| Sun-Joo Cho | 4 |
| A. Corinne Huggins-Manley | 3 |
| Gongjun Xu | 3 |
| Keerat Grewal | 3 |
| Matthew Moreno | 3 |
| Amber Benedict | 2 |
| Diane E. MacKenzie | 2 |
| More ▼ | |
Publication Type
Education Level
Audience
| Teachers | 2 |
| Practitioners | 1 |
| Researchers | 1 |
Location
| Canada | 5 |
| Germany | 2 |
| Israel | 2 |
| Argentina | 1 |
| Australia | 1 |
| China | 1 |
| China (Guangzhou) | 1 |
| Finland | 1 |
| France | 1 |
| Hong Kong | 1 |
| Japan | 1 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
| Program for International… | 3 |
| Trends in International… | 3 |
| Big Five Inventory | 1 |
| National Longitudinal Study… | 1 |
| Stroop Color Word Test | 1 |
What Works Clearinghouse Rating
Yue Liu; Zhen Li; Hongyun Liu; Xiaofeng You – Applied Measurement in Education, 2024
Low test-taking effort of examinees has been considered a source of construct-irrelevant variance in item response modeling, leading to serious consequences on parameter estimation. This study aims to investigate how non-effortful response (NER) influences the estimation of item and person parameters in item-pool scale linking (IPSL) and whether…
Descriptors: Item Response Theory, Computation, Simulation, Responses
Chunyan Liu; Raja Subhiyah; Richard A. Feinberg – Applied Measurement in Education, 2024
Mixed-format tests that include both multiple-choice (MC) and constructed-response (CR) items have become widely used in many large-scale assessments. When an item response theory (IRT) model is used to score a mixed-format test, the unidimensionality assumption may be violated if the CR items measure a different construct from that measured by MC…
Descriptors: Test Format, Response Style (Tests), Multiple Choice Tests, Item Response Theory
Ö. Emre C. Alagöz; Thorsten Meiser – Educational and Psychological Measurement, 2024
To improve the validity of self-report measures, researchers should control for response style (RS) effects, which can be achieved with IRTree models. A traditional IRTree model considers a response as a combination of distinct decision-making processes, where the substantive trait affects the decision on response direction, while decisions about…
Descriptors: Item Response Theory, Validity, Self Evaluation (Individuals), Decision Making
Hanke Vermeiren; Abe D. Hofman; Maria Bolsinova – International Educational Data Mining Society, 2025
The traditional Elo rating system (ERS), widely used as a student model in adaptive learning systems, assumes unidimensionality (i.e., all items measure a single ability or skill), limiting its ability to handle multidimensional data common in educational contexts. In response, several multidimensional extensions of the Elo rating system have been…
Descriptors: Item Response Theory, Models, Comparative Analysis, Algorithms
Kyle T. Turner; George Engelhard Jr. – Journal of Experimental Education, 2024
The purpose of this study is to demonstrate clustering methods within a functional data analysis (FDA) framework for identifying subgroups of individuals that may be exhibiting categories of misfit. Person response functions (PRFs) estimated within a FDA framework (FDA-PRFs) provide graphical displays that can aid in the identification of persons…
Descriptors: Data Analysis, Multivariate Analysis, Individual Characteristics, Behavior
Wan, Siyu; Keller, Lisa A. – Practical Assessment, Research & Evaluation, 2023
Statistical process control (SPC) charts have been widely used in the field of educational measurement. The cumulative sum (CUSUM) is an established SPC method to detect aberrant responses for educational assessments. There are many studies that investigated the performance of CUSUM in different test settings. This paper describes the CUSUM…
Descriptors: Visual Aids, Educational Assessment, Evaluation Methods, Item Response Theory
Zeyuan Jing – ProQuest LLC, 2023
This dissertation presents a comprehensive review of the evolution of DIF analysis within educational measurement from the 1980s to the present. The review elucidates the concept of DIF, particularly emphasizing the crucial role of grouping for exhibiting DIF. Then, the dissertation introduces an innovative modification to the newly developed…
Descriptors: Item Response Theory, Algorithms, Measurement, Test Bias
Ye Ma; Deborah J. Harris – Educational Measurement: Issues and Practice, 2025
Item position effect (IPE) refers to situations where an item performs differently when it is administered in different positions on a test. The majority of previous research studies have focused on investigating IPE under linear testing. There is a lack of IPE research under adaptive testing. In addition, the existence of IPE might violate Item…
Descriptors: Computer Assisted Testing, Adaptive Testing, Item Response Theory, Test Items
Jean-Paul Fox – Journal of Educational and Behavioral Statistics, 2025
Popular item response theory (IRT) models are considered complex, mainly due to the inclusion of a random factor variable (latent variable). The random factor variable represents the incidental parameter problem since the number of parameters increases when including data of new persons. Therefore, IRT models require a specific estimation method…
Descriptors: Sample Size, Item Response Theory, Accuracy, Bayesian Statistics
Sohee Kim; Ki Lynn Cole – International Journal of Testing, 2025
This study conducted a comprehensive comparison of Item Response Theory (IRT) linking methods applied to a bifactor model, examining their performance on both multiple choice (MC) and mixed format tests within the common item nonequivalent group design framework. Four distinct multidimensional IRT linking approaches were explored, consisting of…
Descriptors: Item Response Theory, Comparative Analysis, Models, Item Analysis
Daniel Sánchez; Rachelle Esterhazy – Assessment & Evaluation in Higher Education, 2025
Observational feedback refers to feedback in learning settings that requires observations of student actions. Existing feedback models provide limited insight into the way teachers generate such observational feedback. Addressing this gap, this exploratory study investigates how experienced nursing teachers generate observational feedback in…
Descriptors: Feedback (Response), Observation, Student Evaluation, Nursing Education
Huang, Hung-Yu – Educational and Psychological Measurement, 2023
The forced-choice (FC) item formats used for noncognitive tests typically develop a set of response options that measure different traits and instruct respondents to make judgments among these options in terms of their preference to control the response biases that are commonly observed in normative tests. Diagnostic classification models (DCMs)…
Descriptors: Test Items, Classification, Bayesian Statistics, Decision Making
Stefanie A. Wind; Benjamin Lugu – Applied Measurement in Education, 2024
Researchers who use measurement models for evaluation purposes often select models with stringent requirements, such as Rasch models, which are parametric. Mokken Scale Analysis (MSA) offers a theory-driven nonparametric modeling approach that may be more appropriate for some measurement applications. Researchers have discussed using MSA as a…
Descriptors: Item Response Theory, Data Analysis, Simulation, Nonparametric Statistics
Aiman Mohammad Freihat; Omar Saleh Bani Yassin – Educational Process: International Journal, 2025
Background/purpose: This study aimed to reveal the accuracy of estimation of multiple-choice test items parameters following the models of the item-response theory in measurement. Materials/methods: The researchers depended on the measurement accuracy indicators, which express the absolute difference between the estimated and actual values of the…
Descriptors: Accuracy, Computation, Multiple Choice Tests, Test Items
Hung, Su-Pin; Huang, Hung-Yu – Journal of Educational and Behavioral Statistics, 2022
To address response style or bias in rating scales, forced-choice items are often used to request that respondents rank their attitudes or preferences among a limited set of options. The rating scales used by raters to render judgments on ratees' performance also contribute to rater bias or errors; consequently, forced-choice items have recently…
Descriptors: Evaluation Methods, Rating Scales, Item Analysis, Preferences

Peer reviewed
Direct link
