Publication Date
| In 2026 | 0 |
| Since 2025 | 11 |
| Since 2022 (last 5 years) | 66 |
| Since 2017 (last 10 years) | 144 |
| Since 2007 (last 20 years) | 255 |
Descriptor
| Difficulty Level | 492 |
| Item Analysis | 492 |
| Test Items | 377 |
| Test Construction | 153 |
| Foreign Countries | 118 |
| Multiple Choice Tests | 103 |
| Test Validity | 95 |
| Item Response Theory | 91 |
| Test Reliability | 89 |
| Comparative Analysis | 80 |
| Statistical Analysis | 79 |
| More ▼ | |
Source
Author
| Reckase, Mark D. | 6 |
| Lord, Frederic M. | 5 |
| Roid, Gale | 4 |
| Bratfisch, Oswald | 3 |
| Cahen, Leonard S. | 3 |
| Dorans, Neil J. | 3 |
| Dunne, Tim | 3 |
| Facon, Bruno | 3 |
| Hambleton, Ronald K. | 3 |
| Huck, Schuyler W. | 3 |
| Kostin, Irene | 3 |
| More ▼ | |
Publication Type
Education Level
Audience
| Researchers | 34 |
| Practitioners | 4 |
| Teachers | 2 |
Location
| Indonesia | 8 |
| Nigeria | 8 |
| Turkey | 8 |
| Germany | 7 |
| Taiwan | 7 |
| South Africa | 6 |
| United States | 6 |
| Canada | 5 |
| India | 5 |
| China | 4 |
| Florida | 4 |
| More ▼ | |
Laws, Policies, & Programs
| Education Consolidation… | 1 |
| Elementary and Secondary… | 1 |
| No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Merz, William R.; Grossen, Neal E. – 1978
Six approaches to assessing test item bias were examined: transformed item difficulty, point biserial correlations, chi-square, factor analysis, one parameter item characteristic curve, and three parameter item characteristic curve. Data sets for analysis were generated by a Monte Carlo technique based on the three parameter model; thus, four…
Descriptors: Difficulty Level, Evaluation Methods, Factor Analysis, Item Analysis
Peer reviewedAlderson, J. Charles; Lukmani, Yasmeen – Reading in a Foreign Language, 1989
Attempts to answer the question whether indentifiably separate levels of reading comprehension exist by judgmentally using 9 teachers at the University of Lancaster and empirically using 100 students from Bombay university. (Author/VWL)
Descriptors: Cognitive Ability, College Students, Difficulty Level, English (Second Language)
Peer reviewedPlake, Barbara S.; Melican, Gerald J. – Educational and Psychological Measurement, 1989
The impact of overall test length and difficulty on the expert judgments of item performance by the Nedelsky method were studied. Five university-level instructors predicting the performance of minimally competent candidates on a mathematics examination were fairly consistent in their assessments regardless of length or difficulty of the test.…
Descriptors: Difficulty Level, Estimation (Mathematics), Evaluators, Higher Education
Freedle, Roy; Kostin, Irene – 1991
Analogy items from the Scholastic Aptitude Test (SAT) were evaluated for differential performance by black and white examinees. Black and white examinees were first matched for overall SAT-V scores prior to conducting item analyses. A content and psycholinguistic analysis of 220 disclosed SAT analogy items from 11 test forms was performed.…
Descriptors: Black Students, College Entrance Examinations, Difficulty Level, Ethnic Groups
Abdel-fattah, Abdel-fattah A. – 1992
A scaling procedure is proposed, based on item response theory (IRT), to fit non-hierarchical test structure as well. The binary scores of a test of English were used for calculating the probabilities of answering each item correctly. The probability matrix was factor analyzed, and the difficulty intervals or estimates corresponding to the factors…
Descriptors: Bayesian Statistics, Difficulty Level, English, Estimation (Mathematics)
Higgins, Derrick; Futagi, Yoko; Deane, Paul – ETS Research Report Series, 2005
This paper reports on the process of modifying the ModelCreator item generation system to produce output in multiple languages. In particular, Japanese and Spanish are now supported in addition to English. The addition of multilingual functionality was considerably facilitated by the general formulation of our natural language generation system,…
Descriptors: Multilingualism, Computer Software, Japanese, Spanish
Huntley, Renee M.; Carlson, James E. – 1986
This study compared student performance on language-usage test items presented in two different formats: as discrete sentences and as items embedded in passages. American College Testing (ACT) Program's Assessment experimental units were constructed that presented 40 items in the two different formats. Results suggest item presentation may not…
Descriptors: College Entrance Examinations, Difficulty Level, Goodness of Fit, Item Analysis
Diamond, Esther E. – 1984
The problem of measuring growth across the target grade and, typically, the two adjacent grades, concerns most developers of standardized, norm-referenced achievement tests, particularly at the item selection stage. Opinion is divided on whether to retain or drop items that do not get easier from grade to grade. The controversy has focused on…
Descriptors: Achievement Gains, Achievement Tests, Age Differences, Difficulty Level
Rudner, Lawrence M.; Convey, John J. – 1978
Transformed item difficulties, chi-square, item characteristic curve (icc) theory and factor score techniques were evaluated as approaches for the identification of biased test items. The study was implemented to determine whether the approaches would provide identical classifications of items as to degree of aberrance for culturally different…
Descriptors: Comparative Analysis, Culture Fair Tests, Difficulty Level, Evaluation Criteria
Peer reviewedLyon, Mark A.; Gettinger, Maribeth – Journal of Educational Psychology, 1985
This study examined differences in time needed for learning for three types of learning tasks from Bloom's taxonomy: knowledge, comprehension, and application. Seventh and eighth graders (N=88) were given five repeated trials to master three simulated classroom tasks. Large differences were found between tasks in mastery rates, information…
Descriptors: Comprehension, Content Area Reading, Criterion Referenced Tests, Difficulty Level
Howe, Roger; Scheaffer, Richard; Lindquist, Mary – US Department of Education, 2006
This document contains the framework and a set of recommendations for the NAEP 2007 mathematics assessment, which will assess student achievement nationally and state-by-state, as well as in select urban districts, in grades 4 and 8. It includes descriptions of the mathematical content of the test, the types of test questions, and recommendations…
Descriptors: Grade 4, Mathematical Models, National Competency Tests, Mathematics Instruction
Nenty, H. Johnson – 1986
The Cattell Culture Fair Intelligence Test (CCFIT) was administered to a large sample of American, Nigerian, and Indian adolescents, and item data were examined for cultural bias. The CCFIT was designed to measure fluid intelligence, which is not influenced by cultural differences. Four different item analysis techniques were used to determine…
Descriptors: Construct Validity, Cross Cultural Studies, Cultural Influences, Culture Fair Tests
Mitchell, Karen Janice – 1983
A model of verbal performance was developed which defined cognitive processes thought to underlie performance on the Word Knowledge and Paragraph Comprehension subtests of the Armed Services Vocational Aptitude Battery (ASVAB). The items from two forms of these ASVAB verbal subtests were rated on five conceptualized cognitive storage processes:…
Descriptors: Adults, Aptitude Tests, Cognitive Processes, Difficulty Level
Crocker, Linda; Benson, Jeri – 1977
The initial study was based on responses of 103 examinees on an objective classroom test. Results indicated that response changes increased item difficulties as well as biserial and point biserial correlations. Test reliability was not decreased. This implies that not all students benefit from changing initial responses. The personal biserial…
Descriptors: Achievement Tests, Correlation, Difficulty Level, Graduate Students
Merz, William R.; Rudner, Lawrence M. – 1978
A variety of terms related to test bias or test fairness have been used in a variety of ways, but in this document the "fair use of tests" is defined as equitable selection procedures by means of intact tests, and "test item bias" refers to the study of separate items with respect to the tests of which they are a part. Seven…
Descriptors: Analysis of Covariance, Analysis of Variance, Difficulty Level, Evaluation Criteria


