NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20260
Since 20250
Since 2022 (last 5 years)0
Since 2017 (last 10 years)2
Since 2007 (last 20 years)7
Audience
Researchers2
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 16 to 30 of 53 results Save | Export
Peer reviewed Peer reviewed
Kane, Michael; Moloney, James – Applied Psychological Measurement, 1978
The answer-until-correct (AUC) procedure requires that examinees respond to a multi-choice item until they answer it correctly. Using a modified version of Horst's model for examinee behavior, this paper compares the effect of guessing on item reliability for the AUC procedure and the zero-one scoring procedure. (Author/CTM)
Descriptors: Guessing (Tests), Item Analysis, Mathematical Models, Multiple Choice Tests
Divgi, D. R. – 1980
A method is proposed for providing an absolute, in contrast to comparative, evaluation of how well two tests are equated by transforming their raw scores into a particular common scale. The method is direct, not requiring creation of a standard for comparison; expresses its results in scaled rather than raw scores, and allows examination of the…
Descriptors: Equated Scores, Evaluation Criteria, Item Analysis, Latent Trait Theory
Hendrickson, Gerry F.; Green, Bert F., Jr. – 1972
It has been shown that Guttman weighting of test options results in marked increases in the internal consistency of a test. However, the effect of this type of weighting on the structure of the test is not known. Hence, the purpose of this study is to compare the factor structure of Guttman-weighted and rights-only-weighted tests and to relate the…
Descriptors: Analysis of Variance, Correlation, Factor Analysis, Item Analysis
Jaeger, Richard M. – 1980
Five statistical indices are developed and described which may be used for determining (1) when linear equating of two approximately parallel tests is adequate, and (2) whan a more complex method such as equipercentile equating must be used. The indices were based on: (1) similarity of cumulative score distributions; (2) shape of the raw-score to…
Descriptors: College Entrance Examinations, Difficulty Level, Equated Scores, Higher Education
Veitch, William R. – 1979
The one parameter latent trait theory of Georg Rasch has two assumptions: that student abilities can be measured on an equal interval scale, and that the success of a student with a given item is a function of student achievement and item difficulty. The grade four Michigan Educational Assessment Program reading test was designed to measure…
Descriptors: Cutting Scores, Educational Assessment, Intermediate Grades, Item Analysis
Jones, Bernard G.; Gramenz, Gary W. – Spectrum, 1983
Describes procedure for combining Stanford Achievement Test items with local supplementary items to measure individual and aggregate student performance in mathematics. Two reports are generated from the test results: a diagnostic report of student mastery of each objective and an overall score. (TE)
Descriptors: Criterion Referenced Tests, Elementary Secondary Education, Item Analysis, Item Banks
Ben-Simon, Anat; Bennett, Randy Elliott – Journal of Technology, Learning, and Assessment, 2007
This study evaluated a "substantively driven" method for scoring NAEP writing assessments automatically. The study used variations of an existing commercial program, e-rater[R], to compare the performance of three approaches to automated essay scoring: a "brute-empirical" approach in which variables are selected and weighted solely according to…
Descriptors: Writing Evaluation, Writing Tests, Scoring, Essays
Peer reviewed Peer reviewed
Lord, Frederic M. – Educational and Psychological Measurement, 1971
Descriptors: Ability, Adaptive Testing, Computer Oriented Programs, Difficulty Level
Niemi, David; Wang, Jia; Wang, Haiwen; Vallone, Julia; Griffin, Noelle – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2007
There are usually many testing activities going on in a school, with different tests serving different purposes, thus organization and planning are key in creating an efficient system in assessing the most important educational objectives. In the ideal case, an assessment system will be able to inform on student learning, instruction and…
Descriptors: School Administration, Educational Objectives, Administration, Public Schools
Lowry, Stephen R. – 1979
A specially designed answer format was used for three tests in a college level agriculture class of 19 students to record responses to three things about each item: (1) the student's choice of the best answer; (2) the degree of certainty with which the answer was chosen; and (3) all the answer choices which the student was certain were incorrect.…
Descriptors: Achievement Tests, Confidence Testing, Guessing (Tests), Higher Education
PDF pending restoration PDF pending restoration
Kane, Michael T.; Moloney, James M. – 1976
The Answer-Until-Correct (AUC) procedure has been proposed in order to increase the reliability of multiple-choice items. A model for examinees' behavior when they must respond to each item until they answer it correctly is presented. An expression for the reliability of AUC items, as a function of the characteristics of the item and the scoring…
Descriptors: Guessing (Tests), Item Analysis, Mathematical Models, Multiple Choice Tests
Donlon, Thomas F.; Fitzpatrick, Anne R. – 1978
On the basis of past research efforts to improve multiple-choice test information through differential weighting of responses to wrong answers (distractors), two statistical indices are developed. Each describes the properties of response distributions across the options of an item. Jaspen's polyserial generalization of the biserial correlation…
Descriptors: Confidence Testing, Difficulty Level, Guessing (Tests), High Schools
Sabers, Darrell L.; White, Gordon W. – 1971
A procedure for scoring multiple-choice tests by assigning different weights to every option of a test item is investigated. The weighting method used was based on that proposed by Davis, which involves taking the upper and lower 27% of a sample, according to some criterion measure, and using the percentages of these groups marking an item option…
Descriptors: Computer Oriented Programs, Item Analysis, Measurement Techniques, Multiple Choice Tests
Peer reviewed Peer reviewed
Atkins, Warren J.; And Others – Educational Studies in Mathematics, 1991
Results from the Australian Mathematics Competition for 1988 and 1989 (n=309,443 and n=331,660) were analyzed on three statistical measures to determine risk-taking tendencies by groups of students classified by gender, school year, and achievement level. Results showed statistically significant differences for gender but varied depending on the…
Descriptors: Distractors (Tests), Guessing (Tests), Item Analysis, Mathematics Achievement
Drasgow, Fritz; And Others – 1987
This paper addresses the information revealed in incorrect option selection on multiple choice items. Multilinear Formula Scoring (MFS), a theory providing methods for solving psychological measurement problems of long standing, is first used to estimate option characteristic curves for the Armed Services Vocational Aptitude Battery Arithmetic…
Descriptors: Aptitude Tests, Item Analysis, Latent Trait Theory, Mathematical Models
Pages: 1  |  2  |  3  |  4