Descriptor
| Statistical Analysis | 12 |
| Testing Problems | 12 |
| Test Items | 4 |
| Test Theory | 4 |
| Equated Scores | 3 |
| Item Analysis | 3 |
| Postsecondary Education | 3 |
| Statistical Studies | 3 |
| Test Bias | 3 |
| Test Validity | 3 |
| Comparative Analysis | 2 |
| More ▼ | |
Author
Publication Type
| Speeches/Meeting Papers | 10 |
| Reports - Research | 9 |
| Information Analyses | 1 |
| Journal Articles | 1 |
| Opinion Papers | 1 |
| Reports - Evaluative | 1 |
Education Level
Audience
| Researchers | 12 |
Location
| Canada | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Cope, Ronald T. – 1985
This study considers the use of repeaters when test equating. The subjects consist of five groups of applicants to a professional certification program. Each group comprises first time examinees and repeaters. The procedures include a common item linear equating with nonrandom groups, use of equating chains, and the use of total examinee group…
Descriptors: Certification, Equated Scores, Measurement Techniques, Postsecondary Education
Weber, Larry J.; McBee, Janice K. – 1983
Using multiple choice tests and a statistical method designed to identify flagrant cheaters, the authors undertook to determine (1) the magnitude of cheating on take-home and open-book exams; (2) whether the amount of cheating varied according to three types of examinations (closed-book, open-book or take-home); and (3) if cheating was affected by…
Descriptors: Cheating, College Credits, Higher Education, Multiple Choice Tests
PDF pending restorationvan der Linden, Wim J. – 1984
The classification problem in educational testing is a decision problem. One must assign subjects to one of several available treatments on the basis of test scores, where the success of each treatment is measured by a different criterion. Examples of classification decisions include individualized instruction, counseling, and clinical settings.…
Descriptors: Bayesian Statistics, Classification, Cutting Scores, Decision Making
Peer reviewedKambi, Alan G. – Language, Speech, and Hearing Services in Schools, 1993
This commentary on the mismeasurement of language and reading comprehension abilities argues that quantitative measures of complex behaviors and subsequent ranking of individual performance often do not accurately reflect the abstract constructs they purport to measure, and inappropriate quantification and ranking create and perpetuate potentially…
Descriptors: Elementary Secondary Education, Evaluation Criteria, Evaluation Methods, Evaluation Problems
Doolittle, Allen E. – 1985
Differential item performance (DIP) is discussed as a concept that does not necessarily imply item bias or unfairness to subgroups of examinees. With curriculum-based achievement tests, DIP is presented as a valid reflection of group differences in requisite skills and instruction. Using data from a national testing of the ACT Assessment, this…
Descriptors: Achievement Tests, High Schools, Item Analysis, Mathematics Achievement
Jones, Patricia B.; Sabers, Darrell L. – 1984
Several techniques have been developed for creating continuous smooth distributions of test norms. This paper describes two studies that explore the behavior of cubic splines in order to determine their appropriateness for use in test norming. The first study uses data from the Curriculum Referenced Tests of Mastery (CRTM) and employs two…
Descriptors: Equated Scores, Goodness of Fit, Measurement Techniques, Norm Referenced Tests
Livingston, Samuel A. – 1986
This paper deals with test fairness regarding a test consisting of two parts: (1) a "common" section, taken by all students; and (2) a "variable" section, in which some students may answer a different set of questions from other students. For example, a test taken by several thousand students each year contains a common multiple-choice portion and…
Descriptors: Difficulty Level, Error of Measurement, Essay Tests, Mathematical Models
Bleistein, Carole A. – 1986
Research on assessing the cultural fairness of individual test items is reviewed, with emphasis on Birnbaum's three-parameter logistic model. As defined in this review, differential item characteristics are exhibited when examinees from one group have a lower probability of answering correctly than do examinees of equal ability from another group.…
Descriptors: Black Students, Comparative Analysis, Culture Fair Tests, Literature Reviews
Broussard, Rolland L. – 1985
The cultural bias of the Adult Performance Level Assessment, Form AA-l (APLA) was examined. The potential influence of cultural differences on scores of a major ethnic group, Acadians or Cajuns, was investigated. Assessment items most prone to produce differences in scores were isolated and administered to selected groups. No significant…
Descriptors: Adult Basic Education, Adult Literacy, Culture Fair Tests, Ethnic Groups
Gordon, Howard R. D. – 2001
A random sample of 113 members of the American Vocational Education Research Association (AVERA) was surveyed to obtain baseline information regarding AVERA members' perceptions of statistical significance tests. The Psychometrics Group Instrument was used to collect data from participants. Of those surveyed, 67% were male, 93% had earned a…
Descriptors: Educational Research, Postsecondary Education, Predictor Variables, Research Methodology
Sarvela, Paul D. – 1986
Four discrimination indices were compared, using score distributions which were normal, bimodal, and negatively skewed. The score distributions were systematically varied to represent the common circumstances of a military training situation using criterion-referenced mastery tests. Three 20-item tests were administered to 110 simulated subjects.…
Descriptors: Comparative Analysis, Criterion Referenced Tests, Item Analysis, Mastery Tests
Legg, Sue M.; Algina, James – 1986
This paper focuses on the questions which arise as test practitioners monitor score scales derived from latent trait theory. Large scale assessment programs are dynamic and constantly challenge the assumptions and limits of latent trait models. Even though testing programs evolve, test scores must remain reliable indicators of progress.…
Descriptors: Difficulty Level, Educational Assessment, Elementary Secondary Education, Equated Scores


