Descriptor
Source
Journal of Educational… | 15 |
Author
Publication Type
Journal Articles | 12 |
Reports - Evaluative | 8 |
Reports - Research | 4 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
SAT (College Admission Test) | 2 |
National Assessment of… | 1 |
What Works Clearinghouse Rating

Reckase, Mark D.; And Others – Journal of Educational Measurement, 1988
It is demonstrated, theoretically and empirically, that item sets can be selected that meet the unidimensionality assumption of most item response theory models, even though they require more than one ability for a correct response. A method for identifying such item sets for test development purposes is presented. (SLD)
Descriptors: Computer Simulation, Item Analysis, Latent Trait Theory, Mathematical Models

Barcikowski, Robert S. – Journal of Educational Measurement, 1972
These results indicate that in deciding on the data-gathering design to be used in seeking norm information, attention should be given to item characteristics and test length with particular attention paid to the range of biserial correlations between item response and ability. (Author)
Descriptors: Item Sampling, Mathematical Models, Measurement Techniques, Monte Carlo Methods

Levin, Joel R. – Journal of Educational Measurement, 1975
A set procedure developed in this study is useful in determining sample size, based on specification of linear contrasts involving certain formula treatments. (Author/DEP)
Descriptors: Analysis of Variance, Comparative Analysis, Mathematical Models, Measurement Techniques

Adema, Jos J. – Journal of Educational Measurement, 1990
Mixed integer linear programing models for customizing two-stage tests are presented. Model constraints are imposed with respect to test composition, administration time, inter-item dependencies, and other practical considerations. The models can be modified for use in the construction of multistage tests. (Author/TJH)
Descriptors: Adaptive Testing, Computer Assisted Testing, Equations (Mathematics), Linear Programing

Wainer, Howard; And Others – Journal of Educational Measurement, 1991
A testlet is an integrated group of test items presented as a unit. The concept of testlet differential item functioning (testlet DIF) is defined, and a statistical method is presented to detect testlet DIF. Data from a testlet-based experimental version of the Scholastic Aptitude Test illustrate the methodology. (SLD)
Descriptors: College Entrance Examinations, Definitions, Graphs, Item Bias

Nandakumar, Ratna – Journal of Educational Measurement, 1993
The phenomenon of simultaneous differential item functioning (DIF) amplification and cancellation and the role of the SIBTEST approach in detecting DIF are investigated with a variety of simulated test data. The effectiveness of SIBTEST is supported, and the implications of DIF amplification and cancellation are discussed. (SLD)
Descriptors: Computer Simulation, Elementary Secondary Education, Equal Education, Equations (Mathematics)

Kane, Michael T.; And Others – Journal of Educational Measurement, 1989
This paper develops a multiplicative model as a means of combining ratings of criticality and frequency of various activities involved in job analyses. The model incorporates adjustments to ensure that effective weights of criticality and frequency are appropriate. An example of the model's use is presented. (TJH)
Descriptors: Critical Incidents Method, Higher Education, Job Analysis, Licensing Examinations (Professions)

Hambleton, Ronald K.; De Gruijter, Dato N. M. – Journal of Educational Measurement, 1983
Addressing the shortcomings of classical item statistics for selecting criterion-referenced test items, this paper describes an optimal item selection procedure utilizing item response theory (IRT) and offers examples in which random selection and optimal item selection methods are compared. Theoretical advantages of optimal selection based upon…
Descriptors: Criterion Referenced Tests, Cutting Scores, Item Banks, Latent Trait Theory

Swaminathan, H.; And Others – Journal of Educational Measurement, 1975
A decision-theoretic procedure is outlined which provides a framework within which Bayesian statistical methods can be employed with criterion-referenced tests to improve the quality of decision making in objectives based instructional programs. (Author/DEP)
Descriptors: Bayesian Statistics, Computer Assisted Instruction, Criterion Referenced Tests, Decision Making

Wainer, Howard; Lewis, Charles – Journal of Educational Measurement, 1990
Three different applications of the testlet concept are presented, and the psychometric models most suitable for each application are described. Difficulties that testlets can help overcome include (1) context effects; (2) item ordering; and (3) content balancing. Implications for test construction are discussed. (SLD)
Descriptors: Algorithms, Computer Assisted Testing, Elementary Secondary Education, Item Response Theory

Wainer, Howard; And Others – Journal of Educational Measurement, 1991
Hierarchical (adaptive) and linear methods of testlet construction were compared. The performance of 2,080 ninth and tenth graders on a 4-item testlet was used to predict performance on the entire test. The adaptive test was slightly superior as a predictor, but the cost of obtaining that superiority was considerable. (SLD)
Descriptors: Adaptive Testing, Algebra, Comparative Testing, High School Students

Secolsky, Charles – Journal of Educational Measurement, 1983
A model is presented using examinee judgements in detecting ambiguous/misinterpreted items on teacher-made criterion-referenced tests. A computational example and guidelines for constructing domain categories and interpreting the indices are presented. (Author/PN)
Descriptors: Criterion Referenced Tests, Higher Education, Item Analysis, Mathematical Models

Sireci, Stephen G.; And Others – Journal of Educational Measurement, 1991
Calculating the reliability of a testlet-based test is demonstrated using data from 1,812 males and 2,216 females taking the Scholastic Aptitude Test verbal section and 3,866 examinees taking another reading test. Traditional reliabilities calculated on reading comprehension tests constructed of four testlets provided substantial overestimates.…
Descriptors: College Entrance Examinations, Equations (Mathematics), Estimation (Mathematics), High School Students

Ackerman, Terry A. – Journal of Educational Measurement, 1992
The difference between item bias and item impact and the way they relate to item validity are discussed from a multidimensional item response theory perspective. The Mantel-Haenszel procedure and the Simultaneous Item Bias strategy are used in a Monte Carlo study to illustrate detection of item bias. (SLD)
Descriptors: Causal Models, Computer Simulation, Construct Validity, Equations (Mathematics)

Sheehan, Kathleen; Mislevy, Robert J. – Journal of Educational Measurement, 1990
The 63 items on skills in acquiring and using information from written documents contained in the Survey of Young Adult Literacy in the 1985 National Assessment of Educational Progress are analyzed. The analyses are based on a qualitative cognitive model and an item-response theory model. (TJH)
Descriptors: Adult Literacy, Cognitive Processes, Diagnostic Tests, Elementary Secondary Education