Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 1 |
| Since 2007 (last 20 years) | 5 |
Descriptor
| Comparative Analysis | 8 |
| Guessing (Tests) | 8 |
| Test Format | 8 |
| Multiple Choice Tests | 6 |
| Test Items | 4 |
| Test Reliability | 4 |
| Computer Assisted Testing | 3 |
| Difficulty Level | 2 |
| Foreign Countries | 2 |
| Item Response Theory | 2 |
| Models | 2 |
| More ▼ | |
Source
| Assessment & Evaluation in… | 1 |
| Computers & Education | 1 |
| Educational Psychology | 1 |
| Journal of Educational… | 1 |
| Journal of Intelligence | 1 |
| ProQuest LLC | 1 |
Author
| Baron, Simon | 1 |
| Bernard, David | 1 |
| Bush, Martin | 1 |
| Frary, Robert B. | 1 |
| Jonsson, Bert | 1 |
| Melancon, Janet G. | 1 |
| Myszkowski, Nils | 1 |
| Nicewander, W. Alan | 1 |
| Stenlund, Tova | 1 |
| Stergiopoulos, Charalampos | 1 |
| Storme, Martin | 1 |
| More ▼ | |
Publication Type
| Reports - Research | 7 |
| Journal Articles | 5 |
| Dissertations/Theses -… | 1 |
| Speeches/Meeting Papers | 1 |
Education Level
| Postsecondary Education | 1 |
| Secondary Education | 1 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
| Embedded Figures Test | 1 |
What Works Clearinghouse Rating
Storme, Martin; Myszkowski, Nils; Baron, Simon; Bernard, David – Journal of Intelligence, 2019
Assessing job applicants' general mental ability online poses psychometric challenges due to the necessity of having brief but accurate tests. Recent research (Myszkowski & Storme, 2018) suggests that recovering distractor information through Nested Logit Models (NLM; Suh & Bolt, 2010) increases the reliability of ability estimates in…
Descriptors: Intelligence Tests, Item Response Theory, Comparative Analysis, Test Reliability
Bush, Martin – Assessment & Evaluation in Higher Education, 2015
The humble multiple-choice test is very widely used within education at all levels, but its susceptibility to guesswork makes it a suboptimal assessment tool. The reliability of a multiple-choice test is partly governed by the number of items it contains; however, longer tests are more time consuming to take, and for some subject areas, it can be…
Descriptors: Guessing (Tests), Multiple Choice Tests, Test Format, Test Reliability
Effects of Repeated Testing on Short- and Long-Term Memory Performance across Different Test Formats
Stenlund, Tova; Sundström, Anna; Jonsson, Bert – Educational Psychology, 2016
This study examined whether practice testing with short-answer (SA) items benefits learning over time compared to practice testing with multiple-choice (MC) items, and rereading the material. More specifically, the aim was to test the hypotheses of "retrieval effort" and "transfer appropriate processing" by comparing retention…
Descriptors: Short Term Memory, Long Term Memory, Test Format, Testing
Wu, Yi-Fang – ProQuest LLC, 2015
Item response theory (IRT) uses a family of statistical models for estimating stable characteristics of items and examinees and defining how these characteristics interact in describing item and test performance. With a focus on the three-parameter logistic IRT (Birnbaum, 1968; Lord, 1980) model, the current study examines the accuracy and…
Descriptors: Item Response Theory, Test Items, Accuracy, Computation
Ventouras, Errikos; Triantis, Dimos; Tsiakas, Panagiotis; Stergiopoulos, Charalampos – Computers & Education, 2010
The aim of the present research was to compare the use of multiple-choice questions (MCQs) as an examination method, to the examination based on constructed-response questions (CRQs). Despite that MCQs have an advantage concerning objectivity in the grading process and speed in production of results, they also introduce an error in the final…
Descriptors: Computer Assisted Instruction, Scoring, Grading, Comparative Analysis
Peer reviewedFrary, Robert B. – Journal of Educational Measurement, 1985
Responses to a sample test were simulated for examinees under free-response and multiple-choice formats. Test score sets were correlated with randomly generated sets of unit-normal measures. The extent of superiority of free response tests was sufficiently small so that other considerations might justifiably dictate format choice. (Author/DWH)
Descriptors: Comparative Analysis, Computer Simulation, Essay Tests, Guessing (Tests)
Melancon, Janet G.; Thompson, Bruce – 1988
Applied classical measurement theory was used to study the measurement characteristics of Forms A and B of the Finding Embedded Figures Test (FEFT) when the test is administered in a "no-guessing" or "supply" format. Data provided by 69 students at a private university in the southern United States were used. Both forms of the…
Descriptors: Comparative Analysis, Difficulty Level, Discriminant Analysis, Guessing (Tests)
Nicewander, W. Alan; And Others – 1980
Two methods of interactive, computer-assisted testing methods for multiple-choice items were compared with each other and with conventional multiple-choice tests. The interactive testing methods compared were tailored testing and the respond-until-correct (RUC) item response method. In tailored testing, examinee ability is successively estimated…
Descriptors: Adaptive Testing, Comparative Analysis, Computer Assisted Testing, Guessing (Tests)

Direct link
