Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 3 |
| Since 2007 (last 20 years) | 5 |
Descriptor
| Guessing (Tests) | 13 |
| Test Format | 13 |
| Test Reliability | 13 |
| Multiple Choice Tests | 9 |
| Test Items | 8 |
| Test Validity | 6 |
| Comparative Analysis | 4 |
| Foreign Countries | 4 |
| Higher Education | 4 |
| Scores | 4 |
| Item Analysis | 3 |
| More ▼ | |
Source
Author
| Bush, Martin | 2 |
| Austin, Joe Dan | 1 |
| Baldauf, Richard B., Jr. | 1 |
| Baron, Simon | 1 |
| Bernard, David | 1 |
| Budescu, David V. | 1 |
| Diedenhofen, Birk | 1 |
| Ebel, Robert L. | 1 |
| Frary, Robert B. | 1 |
| Jones, Allan | 1 |
| Joseph, Dane Christian | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 10 |
| Reports - Research | 10 |
| Opinion Papers | 2 |
| Reports - Evaluative | 2 |
| Speeches/Meeting Papers | 2 |
| Dissertations/Theses -… | 1 |
| Guides - Non-Classroom | 1 |
| Tests/Questionnaires | 1 |
Education Level
| Higher Education | 2 |
| Postsecondary Education | 2 |
Audience
| Practitioners | 1 |
| Researchers | 1 |
Location
| France | 1 |
| Germany | 1 |
| United Kingdom | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| Embedded Figures Test | 1 |
What Works Clearinghouse Rating
Papenberg, Martin; Diedenhofen, Birk; Musch, Jochen – Journal of Experimental Education, 2021
Testwiseness may introduce construct-irrelevant variance to multiple-choice test scores. Presenting response options sequentially has been proposed as a potential solution to this problem. In an experimental validation, we determined the psychometric properties of a test based on the sequential presentation of response options. We created a strong…
Descriptors: Test Wiseness, Test Validity, Test Reliability, Multiple Choice Tests
Otoyo, Lucia; Bush, Martin – Practical Assessment, Research & Evaluation, 2018
This article presents the results of an empirical study of "subset selection" tests, which are a generalisation of traditional multiple-choice tests in which test takers are able to express partial knowledge. Similar previous studies have mostly been supportive of subset selection, but the deduction of marks for incorrect responses has…
Descriptors: Multiple Choice Tests, Grading, Test Reliability, Test Format
Storme, Martin; Myszkowski, Nils; Baron, Simon; Bernard, David – Journal of Intelligence, 2019
Assessing job applicants' general mental ability online poses psychometric challenges due to the necessity of having brief but accurate tests. Recent research (Myszkowski & Storme, 2018) suggests that recovering distractor information through Nested Logit Models (NLM; Suh & Bolt, 2010) increases the reliability of ability estimates in…
Descriptors: Intelligence Tests, Item Response Theory, Comparative Analysis, Test Reliability
Bush, Martin – Assessment & Evaluation in Higher Education, 2015
The humble multiple-choice test is very widely used within education at all levels, but its susceptibility to guesswork makes it a suboptimal assessment tool. The reliability of a multiple-choice test is partly governed by the number of items it contains; however, longer tests are more time consuming to take, and for some subject areas, it can be…
Descriptors: Guessing (Tests), Multiple Choice Tests, Test Format, Test Reliability
Joseph, Dane Christian – ProQuest LLC, 2010
Multiple-choice item-writing guideline research is in its infancy. Haladyna (2004) calls for a science of item-writing guideline research. The purpose of this study is to respond to such a call. The purpose of this study was to examine the impact of student ability and method for varying the location of correct answers in classroom multiple-choice…
Descriptors: Evidence, Test Format, Guessing (Tests), Program Effectiveness
Peer reviewedAustin, Joe Dan – Psychometrika, 1981
On distractor-identification tests students mark as many distractors as possible on each test item. A grading scale is developed for this type testing. The score is optimal in that it yields an unbiased estimate of the student's score as if no guessing had occurred. (Author/JKS)
Descriptors: Guessing (Tests), Item Analysis, Measurement Techniques, Scoring Formulas
Ebel, Robert L. – 1981
An alternate-choice test item is a simple declarative sentence, one portion of which is given with two different wordings. For example, "Foundations like Ford and Carnegie tend to be (1) eager (2) hesitant to support innovative solutions to educational problems." The examinee's task is to choose the alternative that makes the sentence…
Descriptors: Comparative Testing, Difficulty Level, Guessing (Tests), Multiple Choice Tests
Peer reviewedKolstad, Rosemarie; And Others – Journal of Dental Education, 1982
Nonrestricted-answer, multiple-choice test items are recommended as a way of including more facts and fewer incorrect answers in test items, and they do not cue successful guessing as restricted multiple choice items can. Examination construction, scoring, and reliability are discussed. (MSE)
Descriptors: Guessing (Tests), Higher Education, Item Analysis, Multiple Choice Tests
Peer reviewedFrary, Robert B. – Journal of Educational Measurement, 1985
Responses to a sample test were simulated for examinees under free-response and multiple-choice formats. Test score sets were correlated with randomly generated sets of unit-normal measures. The extent of superiority of free response tests was sufficiently small so that other considerations might justifiably dictate format choice. (Author/DWH)
Descriptors: Comparative Analysis, Computer Simulation, Essay Tests, Guessing (Tests)
Peer reviewedBaldauf, Richard B., Jr. – Educational and Psychological Measurement, 1982
A Monte Carlo design examined how the effects of guessing and item dependence influence test characteristics and student scores. Although validity for cloze variants was high, multiple-choice cloze had significantly lower reliabilities than did true score equivalents. (Author/PN)
Descriptors: Cloze Procedure, Elementary Education, Guessing (Tests), Reading Comprehension
Melancon, Janet G.; Thompson, Bruce – 1988
Applied classical measurement theory was used to study the measurement characteristics of Forms A and B of the Finding Embedded Figures Test (FEFT) when the test is administered in a "no-guessing" or "supply" format. Data provided by 69 students at a private university in the southern United States were used. Both forms of the…
Descriptors: Comparative Analysis, Difficulty Level, Discriminant Analysis, Guessing (Tests)
Peer reviewedBudescu, David V. – Applied Psychological Measurement, 1988
A multiple matching test--a 24-item Hebrew vocabulary test--was examined, in which distractors from several items are pooled into one list at the test's end. Construction of such tests was feasible. Reliability, validity, and reduction of random guessing were satisfactory when applied to data from 717 applicants to Israeli universities. (SLD)
Descriptors: College Applicants, Feasibility Studies, Foreign Countries, Guessing (Tests)
Peer reviewedJones, Allan – Journal of Geography in Higher Education, 1997
Examines the increase in popularity of objective testing in the United Kingdom and addresses some of the accompanying academic issues. Reports on a case study of test production and implementation to illustrate issues of time costs and benefits. Discusses question styles, marking schemes, and the problem of guesswork. (MJP)
Descriptors: Comparative Testing, Educational Practices, Educational Trends, Foreign Countries

Direct link
