Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 1 |
| Since 2017 (last 10 years) | 1 |
| Since 2007 (last 20 years) | 5 |
Descriptor
| Maximum Likelihood Statistics | 13 |
| Probability | 13 |
| Test Items | 13 |
| Item Response Theory | 7 |
| Mathematical Models | 5 |
| Computation | 4 |
| Simulation | 4 |
| Test Bias | 4 |
| Bayesian Statistics | 3 |
| Classification | 3 |
| Difficulty Level | 3 |
| More ▼ | |
Source
| Journal of Educational and… | 3 |
| ETS Research Report Series | 2 |
| Applied Psychological… | 1 |
| Educational and Psychological… | 1 |
| Measurement in Physical… | 1 |
| Practical Assessment,… | 1 |
| Psychometrika | 1 |
Author
| Wilson, Mark | 2 |
| Andreas Kurz | 1 |
| Baker, Thomas A., III. | 1 |
| Byon, Kevin K. | 1 |
| Can Gürer | 1 |
| Chang, Chih-Hsin | 1 |
| Clemens Draxler | 1 |
| De Boeck, Paul | 1 |
| Futagi, Yoko | 1 |
| Haberman, Shelby J. | 1 |
| Hambleton, Ronald K. | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 9 |
| Reports - Research | 7 |
| Reports - Descriptive | 3 |
| Reports - Evaluative | 2 |
| Numerical/Quantitative Data | 1 |
| Speeches/Meeting Papers | 1 |
Education Level
| Higher Education | 2 |
| Postsecondary Education | 2 |
Audience
Location
| Belgium | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| Graduate Record Examinations | 1 |
What Works Clearinghouse Rating
Clemens Draxler; Andreas Kurz; Can Gürer; Jan Philipp Nolte – Journal of Educational and Behavioral Statistics, 2024
A modified and improved inductive inferential approach to evaluate item discriminations in a conditional maximum likelihood and Rasch modeling framework is suggested. The new approach involves the derivation of four hypothesis tests. It implies a linear restriction of the assumed set of probability distributions in the classical approach that…
Descriptors: Inferences, Test Items, Item Analysis, Maximum Likelihood Statistics
Schuster, Christof; Yuan, Ke-Hai – Journal of Educational and Behavioral Statistics, 2011
Because of response disturbances such as guessing, cheating, or carelessness, item response models often can only approximate the "true" individual response probabilities. As a consequence, maximum-likelihood estimates of ability will be biased. Typically, the nature and extent to which response disturbances are present is unknown, and, therefore,…
Descriptors: Computation, Item Response Theory, Probability, Maximum Likelihood Statistics
Baker, Thomas A., III.; Byon, Kevin K. – Measurement in Physical Education and Exercise Science, 2014
A scale was developed to measure perceptions of sexual abuse in youth sports by assessing (a) the perceived prevalence of sexual abuse committed by pedophilic youth sport coaches, (b) the perceived likelihood that a coach is a pedophile, (c) perceptions on how youth sport organizations should manage the risk of pedophilia, and (d) media influence…
Descriptors: Sexual Abuse, Test Construction, Attitude Measures, Incidence
Paek, Insu; Wilson, Mark – Educational and Psychological Measurement, 2011
This study elaborates the Rasch differential item functioning (DIF) model formulation under the marginal maximum likelihood estimation context. Also, the Rasch DIF model performance was examined and compared with the Mantel-Haenszel (MH) procedure in small sample and short test length conditions through simulations. The theoretically known…
Descriptors: Test Bias, Test Length, Statistical Inference, Geometric Concepts
Rudner, Lawrence M. – Practical Assessment, Research & Evaluation, 2009
This paper describes and evaluates the use of measurement decision theory (MDT) to classify examinees based on their item response patterns. The model has a simple framework that starts with the conditional probabilities of examinees in each category or mastery state responding correctly to each item. The presented evaluation investigates: (1) the…
Descriptors: Classification, Scoring, Item Response Theory, Measurement
Haberman, Shelby J. – ETS Research Report Series, 2006
Multinomial-response models are available that correspond implicitly to tests in which a total score is computed as the sum of polytomous item scores. For these models, joint and conditional estimation may be considered in much the same way as for the Rasch model for right-scored tests. As in the Rasch model, joint estimation is only attractive if…
Descriptors: Computation, Models, Test Items, Scores
Mislevy, Robert J.; Wilson, Mark – 1992
Standard item response theory (IRT) models posit latent variables to account for regularities in students' performance on test items. They can accommodate learning only if the expected changes in performance are smooth, and, in an appropriate metric, uniform over items. Wilson's "Saltus" model extends the ideas of IRT to development that…
Descriptors: Bayesian Statistics, Change, Development, Item Response Theory
Peer reviewedLiou, Michelle; Chang, Chih-Hsin – Psychometrika, 1992
An extension is proposed for the network algorithm introduced by C.R. Mehta and N.R. Patel to construct exact tail probabilities for testing the general hypothesis that item responses are distributed according to the Rasch model. A simulation study indicates the efficiency of the algorithm. (SLD)
Descriptors: Algorithms, Computer Simulation, Difficulty Level, Equations (Mathematics)
Van den Noortgate, Wim; De Boeck, Paul – Journal of Educational and Behavioral Statistics, 2005
Although differential item functioning (DIF) theory traditionally focuses on the behavior of individual items in two (or a few) specific groups, in educational measurement contexts, it is often plausible to regard the set of items as a random sample from a broader category. This article presents logistic mixed models that can be used to model…
Descriptors: Test Bias, Item Response Theory, Educational Assessment, Mathematical Models
Masters, Geoff N.; Wright, Benjamin D. – 1982
The analysis of fit of data to a measurement model for graded responses is described. The model is an extension of Rasch's dichotomous model to formats which provide more than two levels of response to items. The model contains one parameter for each person and one parameter for each "step" in an item. A dichotomously-scored item…
Descriptors: Difficulty Level, Goodness of Fit, Item Analysis, Latent Trait Theory
Peer reviewedSamejima, Fumiko – Applied Psychological Measurement, 1977
The accuracy of estimation of the subjects' latent ability maintained by tailoring for each testee the order of item presentation and the border of item dichotomization was compared to the information provided by the original graded test items. (RC)
Descriptors: Ability, Adaptive Testing, Branching, Computer Assisted Testing
Sheehan, Kathleen M.; Kostin, Irene; Futagi, Yoko; Hemat, Ramin; Zuckerman, Daniel – ETS Research Report Series, 2006
This paper describes the development, implementation, and evaluation of an automated system for predicting the acceptability status of candidate reading-comprehension stimuli extracted from a database of journal and magazine articles. The system uses a combination of classification and regression techniques to predict the probability that a given…
Descriptors: Automation, Prediction, Reading Comprehension, Classification
Hambleton, Ronald K.; And Others – 1977
Latent trait theory supposes that, in testing situations, examinee performance on a test can be predicted (or explained) by defining examinee characteristics, referred to as traits, estimating scores for examinees on these traits and using the scores to predict or explain test performance (Lord and Novick, 1968). In view of the breakthroughs in…
Descriptors: Adaptive Testing, Bayesian Statistics, Cognitive Measurement, Computer Programs

Direct link
