Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 1 |
| Since 2007 (last 20 years) | 6 |
Descriptor
Source
Author
| Bin Usop, Hasbee | 1 |
| Burket, George | 1 |
| Burket, George R. | 1 |
| Burton, Richard F. | 1 |
| Chalkley, Brian | 1 |
| Chen, Li-Sue | 1 |
| Chen, Wen-Hung | 1 |
| Chia, Mike | 1 |
| Clariana, Roy B. | 1 |
| Ercikan, Kadriye | 1 |
| Finnegan, D. J. | 1 |
| More ▼ | |
Publication Type
| Reports - Descriptive | 19 |
| Journal Articles | 12 |
| Speeches/Meeting Papers | 2 |
| Guides - Non-Classroom | 1 |
| Historical Materials | 1 |
| Information Analyses | 1 |
Education Level
| Higher Education | 4 |
| Postsecondary Education | 3 |
| Secondary Education | 2 |
| Elementary Education | 1 |
| Elementary Secondary Education | 1 |
| High Schools | 1 |
Laws, Policies, & Programs
| National Defense Education Act | 1 |
Assessments and Surveys
| National Assessment of… | 2 |
| Sentence Completion Test | 1 |
What Works Clearinghouse Rating
Gorbunova, Tatiana N. – European Journal of Contemporary Education, 2017
The subject of the research is to build methodologies to evaluate the student knowledge by testing. The author points to the importance of feedback about the mastering level in the learning process. Testing is considered as a tool. The object of the study is to create the test system models for defence practice problems. Special attention is paid…
Descriptors: Testing, Evaluation Methods, Feedback (Response), Simulation
McQueen, H. A.; Shields, C.; Finnegan, D. J.; Higham, J.; Simmen, M. W. – Biochemistry and Molecular Biology Education, 2014
We demonstrate that student engagement with PeerWise, an online tool that allows students to author and answer multiple-choice questions (MCQs), is associated with enhanced academic performance across diverse assessment types on a second year Genetics course. Benefits were consistent over three course deliveries, with differential benefits…
Descriptors: Educational Technology, Technology Uses in Education, Science Instruction, Multiple Choice Tests
Wothke, Werner; Burket, George; Chen, Li-Sue; Gao, Furong; Shu, Lianghua; Chia, Mike – Journal of Educational and Behavioral Statistics, 2011
It has been known for some time that item response theory (IRT) models may exhibit a likelihood function of a respondent's ability which may have multiple modes, flat modes, or both. These conditions, often associated with guessing of multiple-choice (MC) questions, can introduce uncertainty and bias to ability estimation by maximum likelihood…
Descriptors: Educational Assessment, Item Response Theory, Computation, Maximum Likelihood Statistics
Sie Hoe, Lau; Ngee Kiong, Lau; Kian Sam, Hong; Bin Usop, Hasbee – Online Submission, 2009
Assessment is central to any educational process. Number Right (NR) scoring method is a conventional scoring method for multiple choice items, where students need to pick one option as the correct answer. One point is awarded for the correct response and zero for any other responses. However, it has been heavily criticized for guessing and failure…
Descriptors: Multiple Choice Tests, Computer Assisted Testing, Adaptive Testing, Scoring
Sawchuk, Stephen – Education Digest: Essential Readings Condensed for Quick Review, 2010
Most experts in the testing community have presumed that the $350 million promised by the U.S. Department of Education to support common assessments would promote those that made greater use of open-ended items capable of measuring higher-order critical-thinking skills. But as measurement experts consider the multitude of possibilities for an…
Descriptors: Educational Quality, Test Items, Comparative Analysis, Multiple Choice Tests
Kurz, Terri Barber – 1999
Multiple-choice tests are generally scored using a conventional number right scoring method. While this method is easy to use, it has several weaknesses. These weaknesses include decreased validity due to guessing and failure to credit partial knowledge. In an attempt to address these weaknesses, psychometricians have developed various scoring…
Descriptors: Algorithms, Guessing (Tests), Item Response Theory, Multiple Choice Tests
Peer reviewedHarper, R. – Journal of Computer Assisted Learning, 2003
Discusses multiple choice questions and presents a statistical approach to post-test correction for guessing that can be used in spreadsheets to automate the correction and generate a grade. Topics include the relationship between the learning objectives and multiple-choice assessments; and guessing correction by negative marking. (LRW)
Descriptors: Behavioral Objectives, Computer Assisted Testing, Grades (Scholastic), Guessing (Tests)
Burton, Richard F. – Assessment and Evaluation in Higher Education, 2005
Examiners seeking guidance on multiple-choice and true/false tests are likely to encounter various faulty or questionable ideas. Twelve of these are discussed in detail, having to do mainly with the effects on test reliability of test length, guessing and scoring method (i.e. number-right scoring or negative marking). Some misunderstandings could…
Descriptors: Guessing (Tests), Multiple Choice Tests, Objective Tests, Test Reliability
Peer reviewedErcikan, Kadriye; Schwartz, Richard D.; Julian, Marc W.; Burket, George R.; Weber, Melba M.; Link, Valerie – Journal of Educational Measurement, 1998
Discusses and demonstrates combining scores from multiple-choice (MC) and constructed-response (CR) items to create a common scale using Item Response Theory methodology. Provides empirical results using a set of tests in reading, language, mathematics, and science in three grades. (SLD)
Descriptors: Constructed Response, Elementary Secondary Education, Item Response Theory, Language Arts
Peer reviewedMunn, R. J.; And Others – Journal of Chemical Education, 1981
Describes a "gradebook program" capable of recording, editing, recoding, weighting, and scaling multiple choice test scores. (SK)
Descriptors: Chemistry, College Science, Computer Assisted Testing, Computer Oriented Programs
Kitao, S. Kathleen; Kitao, Kenji – 1996
The testing of grammar is one of the mainstays of language testing, since it can be argued that a basic knowledge of grammar underlies the ability to use language to express meaning. The most common way of testing grammatical knowledge is the multiple-choice test, and the most common multiple-choice item is one in which the student selects a word…
Descriptors: English (Second Language), Foreign Countries, Grammar, Language Proficiency
Yang, Wen-Ling; Chen, Wen-Hung – 2000
This paper provides a historical review of the changes and improvements made in estimating numerical cutpoints for the National Assessment of Educational Progress (NAEP). While reviewing the various methodologies used for collecting judgment data, the paper discusses: (1) the incorporation of Item Response Theory for setting standards; (2)…
Descriptors: Academic Achievement, Academic Standards, Constructed Response, Cutting Scores
Education Commission of the States, Denver, CO. National Assessment of Educational Progress. – 1981
This handbook describes the procedures used to develop, administer, and analyze the results of the 1978-79 art assessment of 9-year-olds, and 17-year-olds by the National Assessment of Educational Progress (NAEP). The primary purpose of the handbook is to provide detailed procedural information for people interested in replicating the assessment…
Descriptors: Administration, Art Appreciation, Art Education, Art History
Peer reviewedHarton, Helen C.; Green, Laura R.; Jackson, Craig; Latane, Bibb – Teaching of Psychology, 1998
Describes an in-class exercise using students' discussion of their answers on a multiple choice test to demonstrate the powerful effects and pervasiveness of social influence. After a brief multiple choice test, students discuss their answers with their neighbors and then answer again. Similar response patterns emerge among the participants. (MJP)
Descriptors: Cluster Analysis, Correlation, Demonstrations (Educational), Group Behavior
Clariana, Roy B.; Wallace, Patricia – Journal of Educational Computing Research, 2007
This proof-of-concept investigation describes a computer-based approach for deriving the knowledge structure of individuals and of groups from their written essays, and considers the convergent criterion-related validity of the computer-based scores relative to human rater essay scores and multiple-choice test scores. After completing a…
Descriptors: Computer Assisted Testing, Multiple Choice Tests, Construct Validity, Cognitive Structures
Previous Page | Next Page ยป
Pages: 1 | 2
Direct link
