Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 4 |
| Since 2017 (last 10 years) | 5 |
| Since 2007 (last 20 years) | 15 |
Descriptor
Source
Author
Publication Type
Education Level
| Higher Education | 2 |
| Grade 4 | 1 |
| Grade 8 | 1 |
| Postsecondary Education | 1 |
Audience
| Researchers | 49 |
| Practitioners | 1 |
Location
| Netherlands | 5 |
| United States | 3 |
| Australia | 2 |
| Belgium | 2 |
| Italy | 2 |
| California | 1 |
| China | 1 |
| Denmark | 1 |
| Florida | 1 |
| Georgia | 1 |
| Hungary | 1 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Samejima, Fumiko – 1986
Item analysis data fitting the normal ogive model were simulated in order to investigate the problems encountered when applying the three-parameter logistic model. Binary item tests containing 10 and 35 items were created, and Monte Carlo methods simulated the responses of 2,000 and 500 examinees. Item parameters were obtained using Logist 5.…
Descriptors: Computer Simulation, Difficulty Level, Guessing (Tests), Item Analysis
Forster, Fred – 1987
Studies carried out over a 12-year period addressed fundamental questions on the use of Rasch-based item banks. Large field tests administered in grades 3-8 of reading, mathematics, and science items, as well as standardized test results were used to explore the possible effects of many factors on item calibrations. In general, the results…
Descriptors: Achievement Tests, Difficulty Level, Elementary Education, Item Analysis
Downey, Ronald G.
Previous research has studied the effects of different methods of item option weighting on the reliability and concurrent and predictive validity of achievement tests. Increases in reliability are generally found, but with mixed results for validity. Several methods of producing option weights, (i.e., Guttman internal and external weights and…
Descriptors: Achievement Tests, Comparative Analysis, Correlation, Grade Point Average
Lord, Frederic M. – 1971
Some stochastic approximation procedures are considered in relation to the problem of choosing a sequence of test questions to accurately estimate a given examinee's standing on a psychological dimension. Illustrations are given evaluating certain procedures in a specific context. (Author/CK)
Descriptors: Academic Ability, Adaptive Testing, Computer Programs, Difficulty Level
Peer reviewedSamejima, Fumiko – Applied Psychological Measurement, 1977
The accuracy of estimation of the subjects' latent ability maintained by tailoring for each testee the order of item presentation and the border of item dichotomization was compared to the information provided by the original graded test items. (RC)
Descriptors: Ability, Adaptive Testing, Branching, Computer Assisted Testing
Peer reviewedYamamoto, Kentaro; Mazzeo, John – Journal of Educational Statistics, 1992
The need for scale linking in the National Assessment of Educational Progress (NAEP) is discussed, and the specific procedures used to carry out the linking in the context of the major analyses of the 1990 NAEP mathematics assessment are described. Issues remaining to be addressed are outlined. (SLD)
Descriptors: Comparative Testing, Educational Assessment, Elementary Secondary Education, Equated Scores
Peer reviewedWesters, Paul; Kelderman, Henk – Psychometrika, 1992
A method for analyzing test-item responses is proposed to examine differential item functioning (DIF) in multiple-choice items within the latent class framework. Different models for detection of DIF are formulated, defining the subgroup as a latent variable. An efficient estimation method is described and illustrated. (SLD)
Descriptors: Chi Square, Difficulty Level, Educational Testing, Equations (Mathematics)
Samejima, Fumiko – 1990
The shortcomings of the conventional way of using and interpreting multiple-choice tests are summarized. Some theories and methodologies that can be applied for better use multiple-choice test items are described. Empirical facts are introduced to support the theoretical observations. New strategies are proposed that will reduce "noise"…
Descriptors: Ability Identification, Distractors (Tests), Equations (Mathematics), Estimation (Mathematics)
Phillips, Gary W. – 1982
This paper presents an introduction to the use of latent trait models for the estimation of domain scores. It was shown that these models provided an advantage over classical test theory and binomial error models in that unbiased estimates of true domain scores could be obtained even when items were not randomly selected from a universe of items.…
Descriptors: Comparative Analysis, Criterion Referenced Tests, Estimation (Mathematics), Goodness of Fit
Reckase, Mark D. – 1985
Multidimensional item difficulty (MID) is proposed as a means of describing test items which measure more than one ability. With mathematical story problems, for instance, both mathematical and verbal skills are required to obtain a correct answer. The proposed measure of MID is based upon three general assumptions: (1) the probability of…
Descriptors: Ability Identification, College Entrance Examinations, College Mathematics, Difficulty Level
PDF pending restorationReckase, Mark D. – 1986
The work presented in this paper defined conceptually the concepts of multidimensional discrimination and information, derived mathematical expressions for the concepts for a particular multidimensional item response theory (IRT) model, and applied the concepts to actual test data. Multidimensional discrimination was defined as a function of the…
Descriptors: College Entrance Examinations, Difficulty Level, Discriminant Analysis, Item Analysis
McKinley, Robert L.; Reckase, Mark D. – 1981
A study was conducted to compare tailored testing procedures based on a Bayesian ability estimation technique and on a maximum likelihood ability estimation technique. The Bayesian tailored testing procedure selected items so as to minimize the posterior variance of the ability estimate distribution, while the maximum likelihood tailored testing…
Descriptors: Academic Ability, Adaptive Testing, Bayesian Statistics, Comparative Analysis
Curry, Allen R.; And Others – 1978
The efficacy of employing subsets of items from a calibrated item pool to estimate the Rasch model person parameters was investigated. Specifically, the degree of invariance of Rasch model ability-parameter estimates was examined across differing collections of simulated items. The ability-parameter estimates were obtained from a simulation of…
Descriptors: Career Development, Difficulty Level, Equated Scores, Error of Measurement
Yen, Wendy M. – 1979
Three test-analysis models were used to analyze three types of simulated test score data plus the results of eight achievement tests. Chi-square goodness-of-fit statistics were used to evaluate the appropriateness of the models to the four kinds of data. Data were generated to simulate the responses of 1,000 students to 36 pseudo-items by…
Descriptors: Achievement Tests, Correlation, Goodness of Fit, Item Analysis
A Comparison of Three Types of Test Development Procedures Using Classical and Latent Trait Methods.
Benson, Jeri; Wilson, Michael – 1979
Three methods of item selection were used to select sets of 38 items from a 50-item verbal analogies test and the resulting item sets were compared for internal consistency, standard errors of measurement, item difficulty, biserial item-test correlations, and relative efficiency. Three groups of 1,500 cases each were used for item selection. First…
Descriptors: Comparative Analysis, Difficulty Level, Efficiency, Error of Measurement


