NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20260
Since 20250
Since 2022 (last 5 years)0
Since 2017 (last 10 years)1
Since 2007 (last 20 years)7
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of…1
What Works Clearinghouse Rating
Showing 1 to 15 of 24 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Quaigrain, Kennedy; Arhin, Ato Kwamina – Cogent Education, 2017
Item analysis is essential in improving items which will be used again in later tests; it can also be used to eliminate misleading items in a test. The study focused on item and test quality and explored the relationship between difficulty index (p-value) and discrimination index (DI) with distractor efficiency (DE). The study was conducted among…
Descriptors: Item Analysis, Teacher Developed Materials, Test Reliability, Educational Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Ling, Guangming – International Journal of Testing, 2016
To investigate possible iPad related mode effect, we tested 403 8th graders in Indiana, Maryland, and New Jersey under three mode conditions through random assignment: a desktop computer, an iPad alone, and an iPad with an external keyboard. All students had used an iPad or computer for six months or longer. The 2-hour test included reading, math,…
Descriptors: Educational Testing, Computer Assisted Testing, Handheld Devices, Computers
Lai, Cheng-Fei; Irvin, P. Shawn; Park, Bitnara Jasmine; Alonzo, Julie; Tindal, Gerald – Behavioral Research and Teaching, 2012
In this technical report, we present the results of a reliability study of the third-grade multiple choice reading comprehension measures available on the easyCBM learning system conducted in the spring of 2011. Analyses include split-half reliability, alternate form reliability, person and item reliability as derived from Rasch analysis,…
Descriptors: Grade 3, Curriculum Based Assessment, Educational Testing, Testing Programs
Park, Bitnara Jasmine; Irvin, P. Shawn; Lai, Cheng-Fei; Alonzo, Julie; Tindal, Gerald – Behavioral Research and Teaching, 2012
In this technical report, we present the results of a reliability study of the fifth-grade multiple choice reading comprehension measures available on the easyCBM learning system conducted in the spring of 2011. Analyses include split-half reliability, alternate form reliability, person and item reliability as derived from Rasch analysis,…
Descriptors: Grade 5, Curriculum Based Assessment, Educational Testing, Testing Programs
Park, Bitnara Jasmine; Irvin, P. Shawn; Alonzo, Julie; Lai, Cheng-Fei; Tindal, Gerald – Behavioral Research and Teaching, 2012
In this technical report, we present the results of a reliability study of the fourth-grade multiple choice reading comprehension measures available on the easyCBM learning system conducted in the spring of 2011. Analyses include split-half reliability, alternate form reliability, person and item reliability as derived from Rasch analysis,…
Descriptors: Grade 4, Curriculum Based Assessment, Educational Testing, Testing Programs
Irvin, P. Shawn; Alonzo, Julie; Park, Bitnara Jasmine; Lai, Cheng-Fei; Tindal, Gerald – Behavioral Research and Teaching, 2012
In this technical report, we present the results of a reliability study of the sixth-grade multiple choice reading comprehension measures available on the easyCBM learning system conducted in the spring of 2011. Analyses include split-half reliability, alternate form reliability, person and item reliability as derived from Rasch analysis,…
Descriptors: Grade 6, Grade 3, Curriculum Based Assessment, Educational Testing
Joseph, Dane Christian – ProQuest LLC, 2010
Multiple-choice item-writing guideline research is in its infancy. Haladyna (2004) calls for a science of item-writing guideline research. The purpose of this study is to respond to such a call. The purpose of this study was to examine the impact of student ability and method for varying the location of correct answers in classroom multiple-choice…
Descriptors: Evidence, Test Format, Guessing (Tests), Program Effectiveness
Yelvington, James Yowell; Brady, Raymond G. – Educational Technology, 1979
A study of corrective feedback testing, in which scoring is based upon the number of attempts needed to achieve the correct answer, is compared to single attempt testing. (RAO)
Descriptors: Educational Testing, Feedback, Intermode Differences, Multiple Choice Tests
Peer reviewed Peer reviewed
Jessell, John C.; Sullins, Walter L. – Journal of Educational Measurement, 1975
Seven forms of a sixty item multiple choice test were administered to 454 college freshmen for the purpose of studying the effects on reliability and performance of keyed response option sequencing that would be unlikely under a random model. (Author)
Descriptors: College Freshmen, Educational Testing, Multiple Choice Tests, Performance
Peer reviewed Peer reviewed
Ebel, Robert L. – Educational and Psychological Measurement, 1971
Descriptors: Achievement Tests, Educational Testing, Evaluation Methods, Multiple Choice Tests
Garvin, Alfred D.
Confidence weighting (CW) tends to improve the reliability of easy tests; the Coombs-type multiple-response (MR) option tends to improve the reliability of hard tests. It was hypothesized that, on a test of moderate difficulty, offering both the CW and MR response options would improve reliability more than either alone. Twenty-four subjects took…
Descriptors: Confidence Testing, Educational Testing, Multiple Choice Tests, Response Style (Tests)
Garvin, Alfred D.; Ralston, Nancy C. – 1970
Confidence Weighting (CW), after Ebel, and Multiple Responding (MR), after Coombs, are compared empirically to determine which improved test reliability more in the case of a course pretest derived from the final examination. It was hypothesized that MR, which purportedly measures partial knowledge, would be more effective than CW, which measures…
Descriptors: College Students, Educational Testing, Multiple Choice Tests, Objective Tests
Peer reviewed Peer reviewed
Weiten, Wayne – Journal of Experimental Education, 1982
A comparison of double as opposed to single multiple-choice questions yielded significant differences in regard to item difficulty, item discrimination, and internal reliability, but not concurrent validity. (Author/PN)
Descriptors: Difficulty Level, Educational Testing, Higher Education, Multiple Choice Tests
Peer reviewed Peer reviewed
Yelvington, James Yowell; Brady, Raymond G. – Community/Junior College Research Quarterly, 1979
Assesses the applicability of corrective feedback (CF) testing, which allows multiple attempts to respond to a test item, to the community college classroom. Compares CF testing to single answer testing, especially with regard to reliability, equitability, and effect on student motivation. (DD)
Descriptors: Community Colleges, Educational Testing, Feedback, Multiple Choice Tests
Wilcox, Rand R. – 1982
This document contains three papers from the Methodology Project of the Center for the Study of Evaluation. Methods for characterizing test accuracy are reported in the first two papers. "Bounds on the K Out of N Reliability of a Test, and an Exact Test for Hierarchically Related Items" describes and illustrates how an extension of a…
Descriptors: Educational Testing, Evaluation Methods, Guessing (Tests), Latent Trait Theory
Previous Page | Next Page ยป
Pages: 1  |  2