Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 0 |
| Since 2007 (last 20 years) | 1 |
Descriptor
| Item Analysis | 10 |
| Mathematical Formulas | 10 |
| Test Reliability | 10 |
| Research Reports | 4 |
| Test Construction | 4 |
| Test Items | 4 |
| Analysis of Variance | 2 |
| Higher Education | 2 |
| Multiple Choice Tests | 2 |
| Probability | 2 |
| Statistical Analysis | 2 |
| More ▼ | |
Source
| Psychometrika | 2 |
| Educ Psychol Meas | 1 |
| Educational and Psychological… | 1 |
| Journal of Educational… | 1 |
| Journal of Educational and… | 1 |
Author
Publication Type
| Reports - Research | 8 |
| Journal Articles | 5 |
| Speeches/Meeting Papers | 3 |
| Guides - Non-Classroom | 1 |
| Reports - Descriptive | 1 |
| Tests/Questionnaires | 1 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
| SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Longford, Nicholas T. – Journal of Educational and Behavioral Statistics, 2014
A method for medical screening is adapted to differential item functioning (DIF). Its essential elements are explicit declarations of the level of DIF that is acceptable and of the loss function that quantifies the consequences of the two kinds of inappropriate classification of an item. Instead of a single level and a single function, sets of…
Descriptors: Test Items, Test Bias, Simulation, Hypothesis Testing
Peer reviewedJackson, Paul H. – Psychometrika, 1979
Use of the same term "split-half" for division of an n-item test into two subtests containing equal (Cronbach), and possibly unequal (Guttman), numbers of items sometimes leads to a misunderstanding about the relation between Guttman's maximum split-half bound and Cronbach's coefficient alpha. This distinction is clarified. (Author/JKS)
Descriptors: Item Analysis, Mathematical Formulas, Technical Reports, Test Reliability
Gaylord, Richard H. – Educ Psychol Meas, 1969
Descriptors: Correlation, Item Analysis, Mathematical Formulas, Test Construction
Peer reviewedSilverstein, A. B. – Educational and Psychological Measurement, 1980
An alternative derivation was given of Gaylord's formulas showing the relationships among the average item intercorrelation, the average item-test correlation, and test reliability. Certain parallels were also noted in analysis of variance and principal component analysis. (Author)
Descriptors: Analysis of Variance, Item Analysis, Mathematical Formulas, Test Reliability
Wilson, Pamela W.; And Others – 1979
The purpose of this study was to present an empirical correction of the KR21 (Kuder Richardson test reliability) formula that not only yields a closer approximation to the numerical value of the KR20 without overestimation, but also simplifies computation. This correction was accomplished by introducing several correction factors to the numerator…
Descriptors: Higher Education, Item Analysis, Mathematical Formulas, Research Reports
Peer reviewedRaju, Nambury S. – Psychometrika, 1979
An important relationship is given for two generalizations of coefficient alpha: (1) Rajaratnam, Cronbach, and Gleser's generalizability formula for stratified-parallel tests, and (2) Raju's coefficient beta. (Author/CTM)
Descriptors: Item Analysis, Mathematical Formulas, Test Construction, Test Items
Peer reviewedTerwilliger, James S.; Lele, Kaustubh – Journal of Educational Measurement, 1979
Different indices for the internal consistency, reproducibility, or homogeneity of a test are based upon highly similar conceptual frameworks. Illustrations are presented to demonstrate how the maximum and minimum values of KR20 are influenced by test difficulty and the shape of the distribution of test scores. (Author/CTM)
Descriptors: Difficulty Level, Item Analysis, Mathematical Formulas, Statistical Analysis
Brennan, Robert L,; Lockwood, Robert E. – 1979
Procedures for determining cutting scores have been proposed by Angoff and by Nedelsky. Nedelsky's approach requires that a rater examine each distractor within a test item to determine the probability of a minimally competent examinee answering correctly; whereas Angoff uses a judgment based on the whole item, rather than each of its components.…
Descriptors: Achievement Tests, Comparative Analysis, Cutting Scores, Guessing (Tests)
Douglass, James B. – 1979
Undergraduates in five classes were asked to rate their instructors on four general items, and on twenty-four behavior-specific items describing instructor involvement. Three reliability coefficients differing in error score assumptions were calculated for different numbers of students and items, to permit comparison of the reliabilities of both…
Descriptors: Analysis of Variance, Behavior Rating Scales, Higher Education, Item Analysis
Levine, Michael V.; Rubin, Donald B. – 1976
Appropriateness indexes (statistical formulas) for detecting suspiciously high or low scores on aptitude tests were presented, based on a simulation of the Scholastic Aptitude Test (SAT) with 3,000 simulated scores--2,800 normal and 200 suspicious. The traditional index--marginal probability--uses a model for the normal examinee's test-taking…
Descriptors: Academic Ability, Aptitude Tests, College Entrance Examinations, High Schools

Direct link
