Publication Date
| In 2026 | 0 |
| Since 2025 | 55 |
| Since 2022 (last 5 years) | 197 |
| Since 2017 (last 10 years) | 497 |
| Since 2007 (last 20 years) | 745 |
Descriptor
| Test Items | 1189 |
| Test Reliability | 1189 |
| Test Validity | 687 |
| Test Construction | 567 |
| Foreign Countries | 349 |
| Difficulty Level | 280 |
| Item Analysis | 253 |
| Psychometrics | 236 |
| Item Response Theory | 219 |
| Factor Analysis | 184 |
| Multiple Choice Tests | 173 |
| More ▼ | |
Source
Author
| Schoen, Robert C. | 12 |
| LaVenia, Mark | 5 |
| Liu, Ou Lydia | 5 |
| Anderson, Daniel | 4 |
| Bauduin, Charity | 4 |
| DiLuzio, Geneva J. | 4 |
| Farina, Kristy | 4 |
| Haladyna, Thomas M. | 4 |
| Huck, Schuyler W. | 4 |
| Petscher, Yaacov | 4 |
| Stansfield, Charles W. | 4 |
| More ▼ | |
Publication Type
Education Level
Audience
| Practitioners | 39 |
| Researchers | 30 |
| Teachers | 24 |
| Administrators | 13 |
| Support Staff | 3 |
| Counselors | 2 |
| Students | 2 |
| Community | 1 |
| Parents | 1 |
| Policymakers | 1 |
Location
| Turkey | 69 |
| Indonesia | 37 |
| Germany | 20 |
| Canada | 17 |
| Florida | 17 |
| China | 16 |
| Australia | 15 |
| California | 12 |
| Iran | 11 |
| India | 10 |
| New York | 9 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 1 |
| Meets WWC Standards with or without Reservations | 1 |
Loyd, Brenda H. – 1984
One form of adaptive testing involves a two-stage procedure. The first stage is the administration of a routing test. From this first test, an estimate of an examinee's ability is obtained. On the basis of this ability estimate, a second test focused on a given ability level is administered. The purpose of this study was to compare the efficiency…
Descriptors: Academic Ability, Adaptive Testing, Difficulty Level, Elementary Education
Hodgin, Robert F. – 1984
Guidelines for the construction and use of an attitude instrument are presented, and the application of the instrument to measure student attitude toward economics is described. Attention is directed to the Likert-like summated forced-choice variety of attitude instrument, whereby attitude toward the object is inferred from the summed responses to…
Descriptors: Attitude Measures, Economics Education, Higher Education, Item Analysis
Enright, Brian E. – 1982
The paper presents 12 steps in developing and validating criterion referenced tests (CRTs). The author emphasizes the need to closely examine the test's stated purpose and trace the test through the 12 steps in order to find CRTs that are useful rather than useless. Examples are given for each step: preparing or selecting objectives; developing…
Descriptors: Criterion Referenced Tests, Elementary Secondary Education, Test Construction, Test Items
Haenn, Joseph F. – 1981
Procedures for conducting functional level testing have been available for use by practitioners for some time. However, the Title I Evaluation and Reporting System (TIERS), developed in response to the educational amendments of 1974 to the Elementary and Secondary Education Act (ESEA), has provided the impetus for widespread adoption of this…
Descriptors: Achievement Tests, Difficulty Level, Scores, Scoring
Ebel, Robert L. – 1981
An alternate-choice test item is a simple declarative sentence, one portion of which is given with two different wordings. For example, "Foundations like Ford and Carnegie tend to be (1) eager (2) hesitant to support innovative solutions to educational problems." The examinee's task is to choose the alternative that makes the sentence…
Descriptors: Comparative Testing, Difficulty Level, Guessing (Tests), Multiple Choice Tests
Peer reviewedKane, Michael; Moloney, James – Applied Psychological Measurement, 1978
The answer-until-correct (AUC) procedure requires that examinees respond to a multi-choice item until they answer it correctly. Using a modified version of Horst's model for examinee behavior, this paper compares the effect of guessing on item reliability for the AUC procedure and the zero-one scoring procedure. (Author/CTM)
Descriptors: Guessing (Tests), Item Analysis, Mathematical Models, Multiple Choice Tests
Peer reviewedNevo, Barukh – Educational and Psychological Measurement, 1977
Item-test correlations are compared to item test-retest correlations as measures for selecting items in test construction. The author concludes that the item test-retest method is superior for item analysis which aims at getting shorter tests while maintaining test stability. (Author/JKS)
Descriptors: College Students, Correlation, Higher Education, Item Analysis
Peer reviewedDiamond, James J.; McCormick, Janet – Evaluation and the Health Professions, 1986
Using item responses from an in-training examination in diagnostic radiology, the application of a strength of association statistic to the general problem of item analysis is illustrated. Criteria for item selection, general issues of reliability, and error of measurement are discussed. (Author/LMO)
Descriptors: Achievement Tests, Difficulty Level, Error of Measurement, Graduate Medical Education
Peer reviewedHuynh, Huynh – Journal of Educational Statistics, 1986
Under the assumptions of classical measurement theory and the condition of normality, a formula is derived for the reliability of composite scores. The formula represents an extension of the Spearman-Brown formula to the case of truncated data. (Author/JAZ)
Descriptors: Computer Simulation, Error of Measurement, Expectancy Tables, Scoring Formulas
Peer reviewedFeldt, Leonard S. – Educational and Psychological Measurement, 1984
The binomial error model includes form-to-form difficulty differences as error variance and leads to Ruder-Richardson formula 21 as an estimate of reliability. If the form-to-form component is removed from the estimate of error variance, the binomial model leads to KR 20 as the reliability estimate. (Author/BW)
Descriptors: Achievement Tests, Difficulty Level, Error of Measurement, Mathematical Formulas
McCowan, Richard J. – Online Submission, 1999
Item writing is a major responsibility of trainers. Too often, qualified staff who prepare lessons carefully and teach conscientiously use inadequate tests that do not validly reflect the true level of trainee achievement. This monograph describes techniques for constructing multiple-choice items that measure student performance accurately. It…
Descriptors: Multiple Choice Tests, Item Analysis, Test Construction, Test Items
Peer reviewedWeiten, Wayne – Journal of Experimental Education, 1982
A comparison of double as opposed to single multiple-choice questions yielded significant differences in regard to item difficulty, item discrimination, and internal reliability, but not concurrent validity. (Author/PN)
Descriptors: Difficulty Level, Educational Testing, Higher Education, Multiple Choice Tests
Peer reviewedMeredith, Gerald M. – Perceptual and Motor Skills, 1982
The School of Architecture faculty posed the methodological problem to construct a scale of 10 items or less to reliably evaluate instruction at different levels of technical and artistic instruction. Among the first 10 ordered items were: "The instructor did a good job" and "The course was worthwhile." (CM)
Descriptors: Architectural Education, Factor Analysis, Higher Education, Student Evaluation of Teacher Performance
Peer reviewedKolstad, Rosemarie; And Others – Journal of Dental Education, 1982
Nonrestricted-answer, multiple-choice test items are recommended as a way of including more facts and fewer incorrect answers in test items, and they do not cue successful guessing as restricted multiple choice items can. Examination construction, scoring, and reliability are discussed. (MSE)
Descriptors: Guessing (Tests), Higher Education, Item Analysis, Multiple Choice Tests
Peer reviewedPopham, W. James – Reading Horizons, 1982
Details the steps followed in the development of the Basic Skills Word List. (FL)
Descriptors: Elementary Education, Readability, Reading Tests, Test Construction


