Publication Date
| In 2026 | 0 |
| Since 2025 | 3 |
| Since 2022 (last 5 years) | 28 |
| Since 2017 (last 10 years) | 58 |
| Since 2007 (last 20 years) | 213 |
Descriptor
| Comparative Analysis | 253 |
| Evaluation Methods | 253 |
| Feedback (Response) | 110 |
| Item Response Theory | 82 |
| Foreign Countries | 64 |
| Student Evaluation | 47 |
| Models | 45 |
| Student Attitudes | 37 |
| Teaching Methods | 35 |
| Simulation | 33 |
| Scores | 29 |
| More ▼ | |
Source
Author
| Chun Wang | 3 |
| Berenson, Mark L. | 2 |
| Chen, Hanwei | 2 |
| Cohen, Allan S. | 2 |
| Cui, Zhongmin | 2 |
| Finch, Holmes | 2 |
| Gongjun Xu | 2 |
| He, Yong | 2 |
| Humphry, Stephen M. | 2 |
| Huxham, Mark | 2 |
| Kim, Seock-Ho | 2 |
| More ▼ | |
Publication Type
Education Level
Audience
| Researchers | 4 |
| Practitioners | 3 |
| Teachers | 2 |
Location
| Australia | 10 |
| Germany | 6 |
| United Kingdom | 6 |
| Canada | 4 |
| China | 4 |
| Iran | 4 |
| Pennsylvania | 4 |
| Greece | 3 |
| Malaysia | 3 |
| Spain | 3 |
| Taiwan | 3 |
| More ▼ | |
Laws, Policies, & Programs
| Elementary and Secondary… | 1 |
| Social Security | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Markus T. Jansen; Ralf Schulze – Educational and Psychological Measurement, 2024
Thurstonian forced-choice modeling is considered to be a powerful new tool to estimate item and person parameters while simultaneously testing the model fit. This assessment approach is associated with the aim of reducing faking and other response tendencies that plague traditional self-report trait assessments. As a result of major recent…
Descriptors: Factor Analysis, Models, Item Analysis, Evaluation Methods
Sohee Kim; Ki Lynn Cole – International Journal of Testing, 2025
This study conducted a comprehensive comparison of Item Response Theory (IRT) linking methods applied to a bifactor model, examining their performance on both multiple choice (MC) and mixed format tests within the common item nonequivalent group design framework. Four distinct multidimensional IRT linking approaches were explored, consisting of…
Descriptors: Item Response Theory, Comparative Analysis, Models, Item Analysis
Jiaying Xiao; Chun Wang; Gongjun Xu – Grantee Submission, 2024
Accurate item parameters and standard errors (SEs) are crucial for many multidimensional item response theory (MIRT) applications. A recent study proposed the Gaussian Variational Expectation Maximization (GVEM) algorithm to improve computational efficiency and estimation accuracy (Cho et al., 2021). However, the SE estimation procedure has yet to…
Descriptors: Error of Measurement, Models, Evaluation Methods, Item Analysis
Hussain, Zawar; Cheema, Salman Arif; Hussain, Ishtiaq – Sociological Methods & Research, 2022
This article is about making correction in Tarray, Singh, and Zaizai model and further improving it when stratified random sampling is necessary. This is done by using optional randomized response technique in stratified sampling using a combination of Mangat and Singh, Mangat, and Greenberg et al. models. The suggested model has been studied…
Descriptors: Comparative Analysis, Models, Surveys, Questionnaires
Joakim Wallmark; James O. Ramsay; Juan Li; Marie Wiberg – Journal of Educational and Behavioral Statistics, 2024
Item response theory (IRT) models the relationship between the possible scores on a test item against a test taker's attainment of the latent trait that the item is intended to measure. In this study, we compare two models for tests with polytomously scored items: the optimal scoring (OS) model, a nonparametric IRT model based on the principles of…
Descriptors: Item Response Theory, Test Items, Models, Scoring
Klauth, Bo – ProQuest LLC, 2023
In conducting confirmatory factor analysis with ordered response items, the literature suggests that when the number of responses is five and item skewness (IS) is approximately normal, researchers can employ maximum likelihood with robust standard errors (MLR). However, MLR can yield biased factor loadings (FL) and FL standard errors (FLSE) when…
Descriptors: Item Response Theory, Evaluation Methods, Factor Analysis, Error of Measurement
Han, Yuting; Zhang, Jihong; Jiang, Zhehan; Shi, Dexin – Educational and Psychological Measurement, 2023
In the literature of modern psychometric modeling, mostly related to item response theory (IRT), the fit of model is evaluated through known indices, such as X[superscript 2], M2, and root mean square error of approximation (RMSEA) for absolute assessments as well as Akaike information criterion (AIC), consistent AIC (CAIC), and Bayesian…
Descriptors: Goodness of Fit, Psychometrics, Error of Measurement, Item Response Theory
Xieling Chen; Haoran Xie; Di Zou; Lingling Xu; Fu Lee Wang – Educational Technology & Society, 2025
In massive open online course (MOOC) environments, computer-based analysis of course reviews enables instructors and course designers to develop intervention strategies and improve instruction to support learners' learning. This study aimed to automatically and effectively identify learners' concerned topics within their written reviews. First, we…
Descriptors: Classification, MOOCs, Teaching Skills, Artificial Intelligence
Olanipekun, Oluwaseun L.; Zhao, JuLong; Wang, Rongdong; A. Sedory, Stephen; Singh, Sarjinder – Sociological Methods & Research, 2023
In carrying out surveys involving sensitive characteristics, randomized response models have been considered among the best techniques since they provide the maximum privacy protection to the respondents and procure honest responses. Over the years, researchers have carried out studies on the estimation of proportions of the population possessing…
Descriptors: Correlation, Smoking, Thinking Skills, Health Behavior
Xue Zhang; Chun Wang – Grantee Submission, 2022
Item-level fit analysis not only serves as a complementary check to global fit analysis, it is also essential in scale development because the fit results will guide item revision and/or deletion (Liu & Maydeu-Olivares, 2014). During data collection, missing response data may likely happen due to various reasons. Chi-square-based item fit…
Descriptors: Goodness of Fit, Item Response Theory, Scores, Test Length
Bayesian Adaptive Lasso for the Detection of Differential Item Functioning in Graded Response Models
Na Shan; Ping-Feng Xu – Journal of Educational and Behavioral Statistics, 2025
The detection of differential item functioning (DIF) is important in psychological and behavioral sciences. Standard DIF detection methods perform an item-by-item test iteratively, often assuming that all items except the one under investigation are DIF-free. This article proposes a Bayesian adaptive Lasso method to detect DIF in graded response…
Descriptors: Bayesian Statistics, Item Response Theory, Adolescents, Longitudinal Studies
Ma, Boxuan; Hettiarachchi, Gayan Prasad; Fukui, Sora; Ando, Yuji – International Educational Data Mining Society, 2023
Vocabulary proficiency diagnosis plays an important role in the field of language learning, which aims to identify the level of vocabulary knowledge of a learner through his or her learning process periodically, and can be used to provide personalized materials and feedback in language-learning applications. Traditional approaches are widely…
Descriptors: Vocabulary Development, Second Language Instruction, Second Language Learning, Language Proficiency
Chun Wang; Ruoyi Zhu; Gongjun Xu – Grantee Submission, 2022
Differential item functioning (DIF) analysis refers to procedures that evaluate whether an item's characteristic differs for different groups of persons after controlling for overall differences in performance. DIF is routinely evaluated as a screening step to ensure items behavior the same across groups. Currently, the majority DIF studies focus…
Descriptors: Models, Item Response Theory, Item Analysis, Comparative Analysis
Matthew Jon Balcom – ProQuest LLC, 2023
The American school system is funded, operated, and evaluated based on student academic achievement data that is generated by various assessment tools and grading practices. The most traditional grading practice, norm-referenced grading, assesses student academic achievement on a 100-point scale that is further broken into categorical levels (A-F)…
Descriptors: Grading, Chemistry, Science Instruction, Standards
Leech, Tony; Chambers, Lucy – Research Matters, 2022
Two of the central issues in comparative judgement (CJ), which are perhaps underexplored compared to questions of the method's reliability and technical quality, are "what processes do judges use to make their decisions" and "what features do they focus on when making their decisions?" This article discusses both, in the…
Descriptors: Comparative Analysis, Decision Making, Evaluators, Reliability

Peer reviewed
Direct link
