Descriptor
| Comparative Analysis | 4 |
| Difficulty Level | 4 |
| Statistical Studies | 4 |
| Test Items | 3 |
| Cutting Scores | 2 |
| Item Analysis | 2 |
| Mastery Tests | 2 |
| Mathematical Models | 2 |
| Academic Standards | 1 |
| Achievement Tests | 1 |
| Bayesian Statistics | 1 |
| More ▼ | |
Source
| Journal of Educational… | 1 |
Publication Type
| Reports - Research | 4 |
| Speeches/Meeting Papers | 2 |
| Journal Articles | 1 |
Education Level
Audience
| Researchers | 2 |
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Peer reviewedJansen, Margo G. H. – Journal of Educational Statistics, 1986
In this paper a Bayesian procedure is developed for the simultaneous estimation of the reading ability and difficulty parameters which are assumed to be factors in reading errors by the multiplicative Poisson Model. According to several criteria, the Bayesian estimates are better than comparable maximum likelihood estimates. (Author/JAZ)
Descriptors: Achievement Tests, Bayesian Statistics, Comparative Analysis, Difficulty Level
Mills, Craig N.; Melican, Gerald J. – 1987
The study compares three methods for establishing cut-off scores that effect a compromise between absolute cut-offs based on item difficulty and relative cut-offs based on expected passing rates. Each method coordinates these two types of information differently. The Beuk method obtains judges' estimates of an absolute cut-off and an expected…
Descriptors: Academic Standards, Certification, Comparative Analysis, Cutting Scores
Muraki, Eiji – 1984
The TESTFACT computer program and full-information factor analysis of test items were used in a computer simulation conducted to correct for the guessing effect. Full-information factor analysis also corrects for omitted items. The present version of TESTFACT handles up to five factors and 150 items. A preliminary smoothing of the tetrachoric…
Descriptors: Comparative Analysis, Computer Simulation, Computer Software, Correlation
Hambleton, Ronald K.; And Others – 1987
The study compared two promising item response theory (IRT) item-selection methods, optimal and content-optimal, with two non-IRT item selection methods, random and classical, for use in fixed-length certification exams. The four methods were used to construct 20-item exams from a pool of approximately 250 items taken from a 1985 certification…
Descriptors: Comparative Analysis, Content Validity, Cutting Scores, Difficulty Level


