Descriptor
Author
| Kelderman, Henk | 3 |
| Kromrey, Jeffrey D. | 2 |
| Patience, Wayne M. | 2 |
| Reckase, Mark D. | 2 |
| Bacon, Tina P. | 1 |
| Douglas, Graham A. | 1 |
| Du Bose, Pansy | 1 |
| Gustafsson, Jan-Eric | 1 |
| Huynh, Huynh | 1 |
| Macready, George B. | 1 |
| Samejima, Fumiko | 1 |
| More ▼ | |
Publication Type
| Reports - Research | 6 |
| Speeches/Meeting Papers | 6 |
| Reports - Evaluative | 5 |
| Journal Articles | 2 |
Education Level
Audience
| Researchers | 1 |
Location
| Netherlands | 3 |
| South Carolina | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| Comprehensive Tests of Basic… | 2 |
What Works Clearinghouse Rating
Peer reviewedWright, Benjamin D.; Douglas, Graham A. – Applied Psychological Measurement, 1977
A procedure for obtaining Rasch model estimates of item difficulty and of ability is detailed. The procedure approximates the optimal but difficult to obtain "unconditional" estimates. (JKS)
Descriptors: Item Analysis, Latent Trait Theory, Mathematical Models, Measurement
Peer reviewedHuynh, Huynh; Saunders, Joseph C. – Journal of Educational Measurement, 1980
Single administration (beta-binomial) estimates for the raw agreement index p and the corrected-for-chance kappa index in mastery testing are compared with those based on two test administrations in terms of estimation bias and sampling variability. Bias is about 2.5 percent for p and 10 percent for kappa. (Author/RL)
Descriptors: Comparative Analysis, Error of Measurement, Mastery Tests, Mathematical Models
Peer reviewedKelderman, Henk – Psychometrika, 1989
A method is proposed for the detection of item bias with respect to observed or unobserved subgroups, using a loglinear item response theory model assuming a Rasch model for ability and difficulty. A simulation study was performed with 200 sets of data to check the robustness of the method. (SLD)
Descriptors: Equations (Mathematics), Foreign Countries, Higher Education, Item Response Theory
Kelderman, Henk – 1986
A method is proposed for the detection of item bias with respect to observed or unobserved subgroups. The method uses quasi-loglinear models for the incomplete subgroup x test score x item 1 x ... x item k contingency table. If the subgroup membership is unknown, the models are the incomplete-latent-class models of S. J. Haberman (1979). The…
Descriptors: Foreign Countries, Higher Education, Latent Trait Theory, Mathematical Models
Kelderman, Henk; Macready, George B. – 1988
The use of loglinear latent class models to detect item bias was studied. Purposes of the study were to: (1) develop procedures for use in assessing item bias when the grouping variable with respect to which bias occurs is not observed; (2) develop bias detection procedures that relate to a conceptually different assessed trait--a categorical…
Descriptors: Foreign Countries, Higher Education, Latent Trait Theory, Mathematical Models
Kromrey, Jeffrey D.; Bacon, Tina P. – 1992
A Monte Carlo study was conducted to estimate the small sample standard errors and statistical bias of psychometric statistics commonly used in the analysis of achievement tests. The statistics examined in this research were: (1) the index of item difficulty; (2) the index of item discrimination; (3) the corrected item-total point-biserial…
Descriptors: Achievement Tests, Comparative Analysis, Difficulty Level, Estimation (Mathematics)
Gustafsson, Jan-Eric – 1979
Problems and procedures in assessing and obtaining fit of data to the Rasch model are treated and assumptions embodied in the Rasch model are made explicit. It is concluded that statistical tests are needed which are sensitive to deviations so that more than one item parameter would be needed for each item, and more than one person parameter would…
Descriptors: Ability, Difficulty Level, Goodness of Fit, Item Analysis
Samejima, Fumiko – 1986
Item analysis data fitting the normal ogive model were simulated in order to investigate the problems encountered when applying the three-parameter logistic model. Binary item tests containing 10 and 35 items were created, and Monte Carlo methods simulated the responses of 2,000 and 500 examinees. Item parameters were obtained using Logist 5.…
Descriptors: Computer Simulation, Difficulty Level, Guessing (Tests), Item Analysis
Yen, Wendy M. – 1979
Three test-analysis models were used to analyze three types of simulated test score data plus the results of eight achievement tests. Chi-square goodness-of-fit statistics were used to evaluate the appropriateness of the models to the four kinds of data. Data were generated to simulate the responses of 1,000 students to 36 pseudo-items by…
Descriptors: Achievement Tests, Correlation, Goodness of Fit, Item Analysis
Du Bose, Pansy; Kromrey, Jeffrey D. – 1993
Empirical evidence is presented of the relative efficiency of two potential linkage plans to be used when equivalent test forms are being administered. Equating is a process by which scores on one form of a test are converted to scores on another form of the same test. A Monte Carlo study was conducted to examine equating stability and statistical…
Descriptors: Art Education, Comparative Testing, Computer Simulation, Equated Scores
Patience, Wayne M.; Reckase, Mark D. – 1979
Simulated tailored tests were used to investigate the relationships between characteristics of the item pool and the computer program, and the reliability and bias of the resulting ability estimates. The computer program was varied to provide for various step sizes (differences in difficulty between successive steps) and different acceptance…
Descriptors: Adaptive Testing, Computer Assisted Testing, Computer Programs, Educational Testing
Patience, Wayne M.; Reckase, Mark D. – 1979
An experiment was performed with computer-generated data to investigate some of the operational characteristics of tailored testing as they are related to various provisions of the computer program and item pool. With respect to the computer program, two characteristics were varied: the size of the step of increase or decrease in item difficulty…
Descriptors: Adaptive Testing, Computer Assisted Testing, Difficulty Level, Error of Measurement


