Descriptor
Source
Journal of Educational… | 31 |
Author
Mislevy, Robert J. | 3 |
Wainer, Howard | 3 |
Cohen, Allan S. | 2 |
Spray, Judith A. | 2 |
Ackerman, Terry A. | 1 |
Adema, Jos J. | 1 |
De Gruijter, Dato N. M. | 1 |
Divgi, D. R. | 1 |
Foreman, Dale I. | 1 |
Gressard, Risa P. | 1 |
Haertel, Edward H. | 1 |
More ▼ |
Publication Type
Journal Articles | 31 |
Reports - Evaluative | 16 |
Reports - Research | 15 |
Speeches/Meeting Papers | 2 |
Reports - Descriptive | 1 |
Education Level
Audience
Location
South Carolina | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Comprehensive Tests of Basic… | 2 |
National Assessment of… | 2 |
SAT (College Admission Test) | 2 |
Pre Professional Skills Tests | 1 |
What Works Clearinghouse Rating

Reckase, Mark D.; And Others – Journal of Educational Measurement, 1988
It is demonstrated, theoretically and empirically, that item sets can be selected that meet the unidimensionality assumption of most item response theory models, even though they require more than one ability for a correct response. A method for identifying such item sets for test development purposes is presented. (SLD)
Descriptors: Computer Simulation, Item Analysis, Latent Trait Theory, Mathematical Models

Thissen, David; And Others – Journal of Educational Measurement, 1989
An item response model for multiple-choice items is described and illustrated in item analysis. The model provides parametric and graphical summaries of the performance of each alternative associated with a multiple-choice item. The illustrative application of the model involves a pilot test of mathematics achievement items. (TJH)
Descriptors: Distractors (Tests), Latent Trait Theory, Mathematical Models, Mathematics Tests
Monte Carlo Based Null Distribution for an Alternative Goodness-of-Fit Test Statistic in IRT Models.

Stone, Clement A. – Journal of Educational Measurement, 2000
Describes a goodness-of-fit statistic that considers the imprecision with which ability is estimated and involves constructing item fit tables based on each examinee's posterior distribution of ability, given the likelihood of the response pattern and an assumed marginal ability distribution. Also describes a Monte Carlo resampling procedure to…
Descriptors: Goodness of Fit, Item Response Theory, Mathematical Models, Monte Carlo Methods

Spray, Judith A.; Welch, Catherine J. – Journal of Educational Measurement, 1990
The effect of large, within-examinee item difficulty variability on estimates of the proportion of consistent classification of examinees into mastery categories was studied over 2 test administrations for 100 simulated examinees. The proportion of consistent classifications was adequately estimated using the technique proposed by M. Subkoviak…
Descriptors: Classification, Difficulty Level, Estimation (Mathematics), Item Response Theory

Gressard, Risa P.; Loyd, Brenda H. – Journal of Educational Measurement, 1991
A Monte Carlo study, which simulated 10,000 examinees' responses to four tests, investigated the effect of item stratification on parameter estimation in multiple matrix sampling of achievement data. Practical multiple matrix sampling is based on item stratification by item discrimination and a sampling plan with moderate number of subtests. (SLD)
Descriptors: Achievement Tests, Comparative Testing, Computer Simulation, Estimation (Mathematics)

Divgi, D. R. – Journal of Educational Measurement, 1986
This paper discusses various issues involved in using the Rasch Model with multiple-choice tests and questions the suitability of this model for multiple-choice items. Results of some past studies supporting the model are shown to be irrelevant. The effects of the model's misfit on test equating are demonstrated. (Author JAZ)
Descriptors: Equated Scores, Goodness of Fit, Latent Trait Theory, Mathematical Models

Huynh, Huynh; Saunders, Joseph C. – Journal of Educational Measurement, 1980
Single administration (beta-binomial) estimates for the raw agreement index p and the corrected-for-chance kappa index in mastery testing are compared with those based on two test administrations in terms of estimation bias and sampling variability. Bias is about 2.5 percent for p and 10 percent for kappa. (Author/RL)
Descriptors: Comparative Analysis, Error of Measurement, Mastery Tests, Mathematical Models

Adema, Jos J. – Journal of Educational Measurement, 1990
Mixed integer linear programing models for customizing two-stage tests are presented. Model constraints are imposed with respect to test composition, administration time, inter-item dependencies, and other practical considerations. The models can be modified for use in the construction of multistage tests. (Author/TJH)
Descriptors: Adaptive Testing, Computer Assisted Testing, Equations (Mathematics), Linear Programing

Wilcox, Rand R.; And Others – Journal of Educational Measurement, 1988
The second response conditional probability model of decision-making strategies used by examinees answering multiple choice test items was revised. Increasing the number of distractors or providing distractors giving examinees (N=106) the option to follow the model improved results and gave a good fit to data for 29 of 30 items. (SLD)
Descriptors: Cognitive Tests, Decision Making, Mathematical Models, Multiple Choice Tests

Wainer, Howard; And Others – Journal of Educational Measurement, 1991
A testlet is an integrated group of test items presented as a unit. The concept of testlet differential item functioning (testlet DIF) is defined, and a statistical method is presented to detect testlet DIF. Data from a testlet-based experimental version of the Scholastic Aptitude Test illustrate the methodology. (SLD)
Descriptors: College Entrance Examinations, Definitions, Graphs, Item Bias

Nandakumar, Ratna – Journal of Educational Measurement, 1993
The phenomenon of simultaneous differential item functioning (DIF) amplification and cancellation and the role of the SIBTEST approach in detecting DIF are investigated with a variety of simulated test data. The effectiveness of SIBTEST is supported, and the implications of DIF amplification and cancellation are discussed. (SLD)
Descriptors: Computer Simulation, Elementary Secondary Education, Equal Education, Equations (Mathematics)

Veale, James R.; Foreman, Dale I. – Journal of Educational Measurement, 1983
Statistical procedures for measuring heterogeneity of test item distractor distributions, or cultural variation, are presented. These procedures are based on the notion that examinees' responses to the incorrect options of a multiple-choice test provide more information concerning cultural bias than their correct responses. (Author/PN)
Descriptors: Ethnic Bias, Item Analysis, Mathematical Models, Multiple Choice Tests

Masters, Geofferey N. – Journal of Educational Measurement, 1984
This paper develops and illustrates a latent trait approach to constructing an item bank when responses are scored in several ordered categories. This approach is an extension of the methodology developed by Choppin, Wright and Stone, and Wright and Bell for the construction and maintenance of banks of dichotomously scored items. (Author/PN)
Descriptors: Equated Scores, Item Banks, Latent Trait Theory, Mathematical Models

Van der Linden, Wim J. – Journal of Educational Measurement, 1982
An ignored aspect of standard setting, namely the possibility that Angoff or Nedelsky judges specify inconsistent probabilities (e.g., low probabilities for easy items but large probabilities for hard items) is explored. A latent trait method is proposed to estimate such misspecifications, and an index of consistency is defined. (Author/PN)
Descriptors: Cutting Scores, Latent Trait Theory, Mastery Tests, Mathematical Models

Miller, Timothy R.; Spray, Judith A. – Journal of Educational Measurement, 1993
Presents logistic discriminant analysis as a means of detecting differential item functioning (DIF) in items that are polytomously scored. Provides examples of DIF detection using a 27-item mathematics test with 1,977 examinees. The proposed method is simpler and more practical than polytomous extensions of the logistic regression DIF procedure.…
Descriptors: Discriminant Analysis, Item Bias, Mathematical Models, Mathematics Tests