Publication Date
| In 2026 | 0 |
| Since 2025 | 9 |
| Since 2022 (last 5 years) | 50 |
| Since 2017 (last 10 years) | 103 |
| Since 2007 (last 20 years) | 160 |
Descriptor
Source
Author
| Plake, Barbara S. | 7 |
| Huntley, Renee M. | 5 |
| Tollefson, Nona | 4 |
| Wainer, Howard | 4 |
| Baghaei, Purya | 3 |
| Bennett, Randy Elliot | 3 |
| Halpin, Glennelle | 3 |
| Katz, Irvin R. | 3 |
| Lunz, Mary E. | 3 |
| Allen, Nancy L. | 2 |
| Anderson, Paul S. | 2 |
| More ▼ | |
Publication Type
Education Level
Audience
| Researchers | 8 |
| Policymakers | 1 |
| Practitioners | 1 |
| Teachers | 1 |
Location
| Germany | 8 |
| Turkey | 8 |
| Australia | 5 |
| China | 4 |
| Indonesia | 4 |
| Iran | 4 |
| United Kingdom (England) | 4 |
| Canada | 3 |
| Japan | 3 |
| Netherlands | 3 |
| Taiwan | 3 |
| More ▼ | |
Laws, Policies, & Programs
| Pell Grant Program | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Tollefson, Nona; Tripp, Alice – 1986
The item difficulty and item discrimination of three multiple-choice item formats were compared in experimental and non-experimental settings. In the experimental study, 104 graduate students were randomly assigned to complete one of three forms of a multiple-choice test: (1) a complex alternative ("none of the above") as the correct answer; (2) a…
Descriptors: Achievement Tests, Difficulty Level, Discriminant Analysis, Graduate Students
Tollefson, Nona; Tripp, Alice – 1983
This study compared the item difficulty and item discrimination of three multiple choice item formats. The multiple choice formats studied were: a complex alternative (none of the above) as the correct answer; a complex alternative as a foil, and the one-correct answer format. One hundred four graduate students were randomly assigned to complete…
Descriptors: Analysis of Variance, Difficulty Level, Graduate Students, Higher Education
Peer reviewedBennett, Randy Elliot; And Others – Journal of Educational Measurement, 1989
Causes of differential item difficulty for blind students taking the braille edition of the Scholastic Aptitude Test's mathematical section were studied. Data for 261 blind students were compared with data for 8,015 non-handicapped students. Results show an association between selected item categories and differential item functioning. (TJH)
Descriptors: Braille, College Entrance Examinations, Comparative Analysis, Difficulty Level
Peer reviewedFoos, Paul W. – Journal of Experimental Education, 1992
Effects of expected form and expected difficulty of a test were examined for 84 college students expecting an easy or difficult multiple-choice or essay examination but taking a combined test. Results support the hypothesis that individuals work harder, rather than reduce their effort, when difficult work is expected. (SLD)
Descriptors: College Students, Difficulty Level, Essay Tests, Expectation
Peer reviewedCrehan, Kevin D.; And Others – Educational and Psychological Measurement, 1993
Studies with 220 college students found that multiple-choice test items with 3 items are more difficult than those with 4 items, and items with the none-of-these option are more difficult than those without this option. Neither format manipulation affected item discrimination. Implications for test construction are discussed. (SLD)
Descriptors: College Students, Comparative Testing, Difficulty Level, Distractors (Tests)
Peer reviewedPonsoda, Vicente; Olea, Julio; Rodriguez, Maria Soledad; Revuelta, Javier – Applied Measurement in Education, 1999
Compared easy and difficult versions of self-adapted tests (SAT) and computerized adapted tests. No significant differences were found among the tests for estimated ability or posttest state anxiety in studies with 187 Spanish high school students, although other significant differences were found. Discusses implications for interpreting test…
Descriptors: Ability, Adaptive Testing, Comparative Analysis, Computer Assisted Testing
Wang, Xiang-bo; And Others – 1993
An increasingly popular test format allows examinees to choose the items they will answer from among a larger set. When examinee choice is allowed fairness requires that the different test forms thus formed be equated for their possible differential difficulty. For this equating to be possible it is necessary to know how well examinees would have…
Descriptors: Adaptive Testing, Advanced Placement, Difficulty Level, Equated Scores
Nissan, Susan; And Others – 1996
One of the item types in the Listening Comprehension section of the Test of English as a Foreign Language (TOEFL) test is the dialogue. Because the dialogue item pool needs to have an appropriate balance of items at a range of difficulty levels, test developers have examined items at various difficulty levels in an attempt to identify their…
Descriptors: Classification, Dialogs (Language), Difficulty Level, English (Second Language)
Cizek, Gregory J. – 1991
A commonly accepted rule for developing equated examinations using the common-items non-equivalent groups (CINEG) design is that items common to the two examinations being equated should be identical. The CINEG design calls for two groups of examinees to respond to a set of common items that is included in two examinations. In practice, this rule…
Descriptors: Certification, Comparative Testing, Difficulty Level, Higher Education
Kubiak, Anna T.; Cowell, William R. – 1990
A procedure used to average several Mantel-Haenszel delta difference values for an item is described and evaluated. The differential item functioning (DIF) procedure used by the Educational Testing Service (ETS) is based on the Mantel-Haenszel statistical technique for studying matched groups. It is standard procedure at ETS to analyze test items…
Descriptors: Difficulty Level, Elementary Secondary Education, Equated Scores, Item Bias
Peer reviewedFrisbie, David A.; Sweeney, Daryl C. – Journal of Educational Measurement, 1982
A 100-item five-choice multiple choice (MC) biology final exam was converted to multiple choice true-false (MTF) form to yield two content-parallel test forms comprised of the two item types. Students found the MTF items easier and preferred MTF over MC; the MTF subtests were more reliable. (Author/GK)
Descriptors: Biology, College Science, Comparative Analysis, Difficulty Level
Peer reviewedStyles, Irene; Andrich, David – Educational and Psychological Measurement, 1993
This paper describes the use of the Rasch model to help implement computerized administration of the standard and advanced forms of Raven's Progressive Matrices (RPM), to compare relative item difficulties, and to convert scores between the standard and advanced forms. The sample consisted of 95 girls and 95 boys in Australia. (SLD)
Descriptors: Adaptive Testing, Computer Assisted Testing, Difficulty Level, Elementary Education
Xu, Yuejin; Iran-Nejad, Asghar; Thoma, Stephen J. – Journal of Interactive Online Learning, 2007
The purpose of the study was to determine comparability of an online version to the original paper-pencil version of Defining Issues Test 2 (DIT2). This study employed methods from both Classical Test Theory (CTT) and Item Response Theory (IRT). Findings from CTT analyses supported the reliability and discriminant validity of both versions.…
Descriptors: Computer Assisted Testing, Test Format, Comparative Analysis, Test Theory
Sykes, Robert C.; Ito, Kyoko – 1995
Whether the presence of bidimensionality has any effect on the adaptive recalibration of test items was studied through live-data simulation of computer adaptive testing (CAT) forms. The source data were examinee responses to the 298 scored multiple choice items of a licensure examination in a health care profession. Three 75-item part-forms,…
Descriptors: Adaptive Testing, Computer Assisted Testing, Difficulty Level, Estimation (Mathematics)
Martinez, Michael E.; Katz, Irvin R. – 1992
Contrasts between constructed response items and stem-equivalent multiple-choice counterparts typically have involved averaging item characteristics, and this aggregation has masked differences in statistical properties at the item level. Moreover, even aggregated format differences have not been explained in terms of differential cognitive…
Descriptors: Architecture, Cognitive Processes, Construct Validity, Constructed Response

Direct link
