NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 316 to 330 of 492 results Save | Export
Peer reviewed Peer reviewed
Lumsden, James – Applied Psychological Measurement, 1977
Person changes can be of three kinds: developmental trends, swells, and tremors. Person unreliability in the tremor sense (momentary fluctuations) can be estimated from person characteristic curves. Average person reliability for groups can be compared from item characteristic curves. (Author)
Descriptors: Difficulty Level, Individual Characteristics, Individual Development, Individual Differences
Peer reviewed Peer reviewed
Lord, Frederic M. – Educational and Psychological Measurement, 1971
A number of empirical studies are suggested to answer certain questions in connection with flexilevel tests. (MS)
Descriptors: Comparative Analysis, Difficulty Level, Guessing (Tests), Item Analysis
Peer reviewed Peer reviewed
Green, Kathy E. – Educational and Psychological Measurement, 1983
This study was concerned with the reliability and validity of subjective judgments about five characteristics of multiple-choice test items from an introductory college-level astronomy test: (1) item difficulty, (2) language complexity, (3) content importance or relevance, (4) response set convergence, and (5) process complexity. (Author)
Descriptors: Achievement Tests, Astronomy, Difficulty Level, Evaluative Thinking
Peer reviewed Peer reviewed
Prien, Borge – Studies in Educational Evaluation, 1989
Under certain conditions it may be possible to determine the difficulty of previously untested test items. Although no recipe can be provided, reflections on this topic are presented, drawing on concepts of item banking. A functional constructive method is suggested as having the most potential. (SLD)
Descriptors: Difficulty Level, Educational Assessment, Foreign Countries, Item Analysis
Peer reviewed Peer reviewed
Henning, Grant – Language Testing, 1988
Violations of item unidimensionality on language tests produced distorted estimates of person ability, and violations of person unidimensionality produced distorted estimates of item difficulty. The Bejar Method was sensitive to such distortions. (Author)
Descriptors: Construct Validity, Content Validity, Difficulty Level, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Sakko, Gina; Martin, Toby L.; Vause, Tricia; Martin, Garry L.; Yu, C. T. – American Journal on Mental Retardation, 2004
The Assessment of Basic Learning Abilities test (ABLA) is a useful tool for choosing appropriate training tasks for persons with developmental disabilities. This test assesses the ease or difficulty with which persons are able to learn six hierarchically positioned discrimination tasks. A visual-visual nonidentity matching prototype task was…
Descriptors: Developmental Disabilities, Learning Problems, Task Analysis, Predictive Validity
Matlock-Hetzel, Susan – 1997
When norm-referenced tests are developed for instructional purposes, to assess the effects of educational programs, or for educational research purposes, it can be very important to conduct item and test analyses. These analyses can evaluate the quality of items and of the test as a whole. Such analyses can also be employed to revise and improve…
Descriptors: Difficulty Level, Distractors (Tests), Elementary Secondary Education, Item Analysis
Tollefson, Nona; Chen, Ju Shan – 1986
This study compared item difficulty and item discrimination indices for parallel multiple-choice items in three content areas: measurement concepts, statistical terminology, and synonyms. The statistics and measurement items were administered in classes where graduate students taking the test were studying the content. Vocabulary items represented…
Descriptors: Difficulty Level, Graduate Students, Higher Education, Item Analysis
PDF pending restoration PDF pending restoration
Cobern, William W. – 1986
This computer program, written in BASIC, performs three different calculations of test reliability: (1) the Kuder-Richardson method; (2); the "common split-half" method; and (3) the Rulon-Guttman split-half method. The program reads sequential access data files for microcomputers that have been set up by statistical packages such as…
Descriptors: Computer Software, Difficulty Level, Educational Research, Equations (Mathematics)
Sinnott, Loraine T. – 1982
A standard method for exploring item bias is the intergroup comparison of item difficulties. This paper describes a refinement and generalization of this technique. In contrast to prior approaches, the proposed method deletes outlying items from the formulation of a criterion for identifying items as deviant. It also extends the mathematical…
Descriptors: College Entrance Examinations, Difficulty Level, Higher Education, Item Analysis
Winsberg, Suzanne; And Others – 1984
In most item response theory models a particular mathematical form is assumed for all item characteristic curves, e.g., a logistic function. It could be desirable, however, to estimate the shape of the item characteristic curves without prior restrictive assumptions about its mathematical form. We have developed a practical method of estimating…
Descriptors: Difficulty Level, Estimation (Mathematics), Goodness of Fit, Item Analysis
Weiten, Wayne – 1979
Two different formats for multiple-choice test items were compared in an experimental test given in a college class in introductory psychology. In one format, a question or incomplete statement was followed by four answers or completions, only one of which was correct. In the other format, the double multiple-choice version, the same questions…
Descriptors: Difficulty Level, Higher Education, Item Analysis, Multiple Choice Tests
Samejima, Fumiko – 1980
Many combinations of a method and an approach for estimating the operating characteristics of the graded item responses, without assuming any mathematical forms, have been produced. In these methods, a set of items whose characteristics are known, or Old Test, is used, which has a large, constant amount of test information throughout the interval…
Descriptors: Difficulty Level, Item Analysis, Latent Trait Theory, Least Squares Statistics
Kolen, Michael J.; Whitney, Douglas R. – 1978
The application of latent trait theory to classroom tests necessitates the use of small sample sizes for parameter estimation. Computer generated data were used to assess the accuracy of estimation of the slope and location parameters in the two parameter logistic model with fixed abilities and varying small sample sizes. The maximum likelihood…
Descriptors: Difficulty Level, Item Analysis, Latent Trait Theory, Mathematical Models
Lenke, Joanne M.; And Others – 1977
To investigate the effect of violating the assumption of equal item difficulty on Kuder-Richardson (KR) Formula 21 reliability coefficient, 670 eighth-and ninth- grade students were administered 26 short, homogeneous "tests" of mathematics concepts and skills. Both KR Formula 20 and KR Formula 21 were used to estimate reliability on each…
Descriptors: Comparative Analysis, Diagnostic Tests, Difficulty Level, Item Analysis
Pages: 1  |  ...  |  18  |  19  |  20  |  21  |  22  |  23  |  24  |  25  |  26  |  ...  |  33