Publication Date
| In 2026 | 0 |
| Since 2025 | 85 |
| Since 2022 (last 5 years) | 453 |
| Since 2017 (last 10 years) | 1241 |
| Since 2007 (last 20 years) | 2515 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 122 |
| Teachers | 105 |
| Researchers | 64 |
| Students | 46 |
| Administrators | 14 |
| Policymakers | 7 |
| Counselors | 3 |
| Parents | 3 |
Location
| Canada | 134 |
| Turkey | 131 |
| Australia | 123 |
| Iran | 66 |
| Indonesia | 61 |
| United Kingdom | 51 |
| Germany | 50 |
| Taiwan | 46 |
| United States | 43 |
| China | 39 |
| California | 35 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 3 |
| Meets WWC Standards with or without Reservations | 5 |
| Does not meet standards | 6 |
Peer reviewedCrehan, Kevin; Haladyna, Thomas M. – Journal of Experimental Education, 1991
Two item-writing rules were tested: phrasing stems as questions versus partial sentences; and using the "none-of-the-above" option instead of a specific content option. Results with 228 college students do not support the use of either stem type and provide limited evidence to caution against the "none-of-the-above" option.…
Descriptors: College Students, Higher Education, Multiple Choice Tests, Test Construction
Peer reviewedPerkins, Kyle; And Others – International Journal of Applied Linguistics, 1991
Discusses a concurrent validity study of an indirect measure of English-as-a-Second-Language writing based on information processing: anagram solving, word reordering, paragraph assembly tests, etc. The criterion measure was a direct, holistic measure of writing. Analysis revealed that the indirect measure did not exhibit concurrent validity for…
Descriptors: English (Second Language), Information Processing, Multiple Choice Tests, Test Validity
Sternberg, Robert J.; Torff, Bruce; Grigorenko, Elena – Phi Delta Kappan, 1998
A "successful intelligence" intervention improved school achievement for a group of 225 ethnically diverse third-graders, both on performance assessments measuring analytical, creative, and practical achievements and on conventional multiple-choice memory assessments. Teaching for triarchic thinking facilitates factual recall, because learning…
Descriptors: Academic Achievement, Grade 3, Intelligence, Intervention
Peer reviewedWang, Wen-chung – Journal of Outcome Measurement, 1998
A Rasch-type analysis is presented for multiple-choice items in which one parameter is assigned to each distractor. Results of a small simulation study show that the parameter recovery of the distractor model is very satisfactory. Analysis of a real dataset shows that some items fit the Rasch model rather than the distractor model. (SLD)
Descriptors: Distractors (Tests), Goodness of Fit, Item Response Theory, Multiple Choice Tests
Peer reviewedHarris, Diana K.; And Others – Educational Gerontology, 1996
Multiple-choice and true-false versions of Palmore's first Facts on Aging Quiz were completed by 501 college students. Multiple choice reduced the chances of guessing and had less measurement error for average and above-average respondents. (Author/SK)
Descriptors: Aging (Individuals), College Students, Error of Measurement, Guessing (Tests)
Ingram, Ella L.; Nelson, Craig E. – American Biology Teacher, 2006
Multiple choice questions are a common method of assessing student understanding. In this article, the authors discuss and evaluate a student-focused collaborative learning strategy for working with such questions that results in greater student learning and allows instructors to better understand student thinking and ultimately to write better…
Descriptors: Multiple Choice Tests, Misconceptions, Cooperative Learning, Teaching Methods
Kim, Seonghoon; Kolen, Michael J. – Applied Measurement in Education, 2006
Four item response theory linking methods (2 moment methods and 2 characteristic curve methods) were compared to concurrent (CO) calibration with the focus on the degree of robustness to format effects (FEs) when applying the methods to multidimensional data that reflected the FEs associated with mixed-format tests. Based on the quantification of…
Descriptors: Item Response Theory, Robustness (Statistics), Test Format, Comparative Analysis
Briggs, Derek C.; Alonzo, Alicia C.; Schwab, Cheryl; Wilson, Mark – Educational Assessment, 2006
In this article we describe the development, analysis, and interpretation of a novel item format we call Ordered Multiple-Choice (OMC). A unique feature of OMC items is that they are linked to a model of student cognitive development for the construct being measured. Each of the possible answer choices in an OMC item is linked to developmental…
Descriptors: Diagnostic Tests, Multiple Choice Tests, Cognitive Development, Item Response Theory
Bush, Martin E. – Quality Assurance in Education: An International Perspective, 2006
Purpose: To provide educationalists with an understanding of the key quality issues relating to multiple-choice tests, and a set of guidelines for the quality assurance of such tests. Design/methodology/approach: The discussion of quality issues is structured to reflect the order in which those issues naturally arise. It covers the design of…
Descriptors: Multiple Choice Tests, Test Reliability, Educational Quality, Quality Control
Hautau, Briana; Turner, Haley C.; Carroll, Erin; Jaspers, Kathryn; Krohn, Katy; Parker, Megan; Williams, Robert L. – Journal of Behavioral Education, 2006
Students (N=153) in three equivalent sections of an undergraduate human development course compared pairs of related concepts via either written or oral discussion at the beginning of most class sessions. A writing-for-random-credit section achieved significantly higher ratings on the writing activities than did a writing-for-no-credit section.…
Descriptors: Writing Exercises, Multiple Choice Tests, Undergraduate Study, Credits
Bleske-Rechek, April; Zeug, Nicole; Webb, Rose Mary – Assessment & Evaluation in Higher Education, 2007
We conducted correlational and performance discrepancy analyses on exam and achievement data taken from students in three psychology courses. Across courses, the same findings emerged. First, only a small fraction of students consistently performed more strongly on one type of assessment (e.g., multiple-choice) than on another (e.g., short…
Descriptors: Psychology, Scores, Academic Aptitude, Academic Achievement
Hartley, James; Betts, Lucy; Murray, Wayne – Psychology Teaching Review, 2007
Background: Recent changes in higher education in the UK have led to much discussion about the performance of men and women students with different methods of assessment. Aim: To see whether or not there were differences between the marks awarded to men and women final-year psychology students as a function of the modes of assessment used. Method:…
Descriptors: Student Evaluation, Females, Psychology, Males
Emurian, Henry H. – Behavior Analyst Today, 2007
At the beginning of a Java computer programming course, nine students in an undergraduate class and nine students in a graduate class completed a web-based programmed instruction tutoring system that taught a simple computer program. All students exited the tutor with an identical level of skill, at least as determined by the tutor's required…
Descriptors: Multiple Choice Tests, Computer Software, Computers, Program Effectiveness
Badgett, John L.; Christmann, Edwin P. – Corwin, 2009
While today's curriculum is largely driven by standards, many teachers find the lack of specificity in the standards to be confounding and even intimidating. Now this practical book provides middle and high school teachers with explicit guidance on designing specific objectives and developing appropriate formative and summative assessments to…
Descriptors: Test Items, Student Evaluation, Knowledge Level, National Standards
Kehoe, Jerard – 1995
This digest presents a list of recommendations for writing multiple-choice test items, based on psychometrics statistics are typically provided by a measurement, or test scoring, service, where tests are machine-scored or by testing software packages. Test makers can capitalize on the fact that "bad" items can be differentiated from…
Descriptors: Item Analysis, Item Banks, Measurement Techniques, Multiple Choice Tests

Direct link
