Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 1 |
| Since 2007 (last 20 years) | 7 |
Descriptor
| Comparative Testing | 14 |
| Item Analysis | 14 |
| Multiple Choice Tests | 14 |
| Test Format | 9 |
| Test Items | 9 |
| Higher Education | 6 |
| Difficulty Level | 5 |
| Test Construction | 4 |
| Test Validity | 4 |
| Mathematics Tests | 3 |
| Performance Based Assessment | 3 |
| More ▼ | |
Source
Author
| Huntley, Renee M. | 2 |
| Kim, Sooyeon | 2 |
| Albanese, Mark A. | 1 |
| Bauer, Daniel | 1 |
| Boughton, Keith | 1 |
| Brice, Julie | 1 |
| Chissom, Brad | 1 |
| Chukabarah, Prince C. O. | 1 |
| Coombes, Lee | 1 |
| Crehan, Kevin D. | 1 |
| Ehmer, Maike | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 10 |
| Reports - Research | 9 |
| Reports - Evaluative | 4 |
| Speeches/Meeting Papers | 4 |
| Reports - Descriptive | 1 |
Education Level
| Elementary Education | 2 |
| Higher Education | 2 |
| Elementary Secondary Education | 1 |
| Grade 4 | 1 |
Audience
| Researchers | 1 |
Location
| Germany | 2 |
Laws, Policies, & Programs
Assessments and Surveys
| ACT Assessment | 2 |
| Embedded Figures Test | 1 |
What Works Clearinghouse Rating
Lahner, Felicitas-Maria; Lörwald, Andrea Carolin; Bauer, Daniel; Nouns, Zineb Miriam; Krebs, René; Guttormsen, Sissel; Fischer, Martin R.; Huwendiek, Sören – Advances in Health Sciences Education, 2018
Multiple true-false (MTF) items are a widely used supplement to the commonly used single-best answer (Type A) multiple choice format. However, an optimal scoring algorithm for MTF items has not yet been established, as existing studies yielded conflicting results. Therefore, this study analyzes two questions: What is the optimal scoring algorithm…
Descriptors: Scoring Formulas, Scoring Rubrics, Objective Tests, Multiple Choice Tests
Laprise, Shari L. – College Teaching, 2012
Successful exam composition can be a difficult task. Exams should not only assess student comprehension, but be learning tools in and of themselves. In a biotechnology course delivered to nonmajors at a business college, objective multiple-choice test questions often require students to choose the exception or "not true" choice. Anecdotal student…
Descriptors: Feedback (Response), Test Items, Multiple Choice Tests, Biotechnology
Sparfeldt, Jorn R.; Kimmel, Rumena; Lowenkamp, Lena; Steingraber, Antje; Rost, Detlef H. – Educational Assessment, 2012
Multiple-choice (MC) reading comprehension test items comprise three components: text passage, questions about the text, and MC answers. The construct validity of this format has been repeatedly criticized. In three between-subjects experiments, fourth graders (N[subscript 1] = 230, N[subscript 2] = 340, N[subscript 3] = 194) worked on three…
Descriptors: Test Items, Reading Comprehension, Construct Validity, Grade 4
Ricketts, Chris; Brice, Julie; Coombes, Lee – Advances in Health Sciences Education, 2010
The purpose of multiple choice tests of medical knowledge is to estimate as accurately as possible a candidate's level of knowledge. However, concern is sometimes expressed that multiple choice tests may also discriminate in undesirable and irrelevant ways, such as between minority ethnic groups or by sex of candidates. There is little literature…
Descriptors: Medical Students, Testing Accommodations, Ethnic Groups, Learning Disabilities
Kim, Sooyeon; Walker, Michael E.; McHale, Frederick – Journal of Educational Measurement, 2010
In this study we examined variations of the nonequivalent groups equating design for tests containing both multiple-choice (MC) and constructed-response (CR) items to determine which design was most effective in producing equivalent scores across the two tests to be equated. Using data from a large-scale exam, this study investigated the use of…
Descriptors: Measures (Individuals), Scoring, Equated Scores, Test Bias
Hammann, Marcus; Phan, Thi Thanh Hoi; Ehmer, Maike; Grimm, Tobias – Journal of Biological Education, 2008
This study is concerned with different forms of assessment of pupils' skills in experimentation. The findings of three studies are reported. Study 1 investigates whether it is possible to develop reliable multiple-choice tests for the skills of forming hypotheses, designing experiments and analysing experimental data. Study 2 compares scores from…
Descriptors: Multiple Choice Tests, Experiments, Science Process Skills, Skill Analysis
Puhan, Gautam; Boughton, Keith; Kim, Sooyeon – Journal of Technology, Learning, and Assessment, 2007
The study evaluated the comparability of two versions of a certification test: a paper-and-pencil test (PPT) and computer-based test (CBT). An effect size measure known as Cohen's d and differential item functioning (DIF) analyses were used as measures of comparability at the test and item levels, respectively. Results indicated that the effect…
Descriptors: Computer Assisted Testing, Effect Size, Test Bias, Mathematics Tests
Peer reviewedCrehan, Kevin D.; And Others – Educational and Psychological Measurement, 1993
Studies with 220 college students found that multiple-choice test items with 3 items are more difficult than those with 4 items, and items with the none-of-these option are more difficult than those without this option. Neither format manipulation affected item discrimination. Implications for test construction are discussed. (SLD)
Descriptors: College Students, Comparative Testing, Difficulty Level, Distractors (Tests)
Huntley, Renee M.; And Others – 1990
This study investigated the effect of diagram formats on performance on geometry items in order to determine whether certain examinees are affected by different item formats and whether such differences arise from the different intellectual demands made by these formats. Thirty-two experimental, multiple-choice geometry items were administered in…
Descriptors: College Bound Students, College Entrance Examinations, Comparative Testing, Diagrams
Chissom, Brad; Chukabarah, Prince C. O. – 1985
The comparative effects of various sequences of test items were examined for over 900 graduate students enrolled in an educational research course at The University of Alabama, Tuscaloosa. experiment, which was conducted a total of four times using four separate tests, presented three different arrangements of 50 multiple-choice items: (1)…
Descriptors: Analysis of Variance, Comparative Testing, Difficulty Level, Graduate Students
Peer reviewedKent, Thomas H.; Albanese, Mark A. – Evaluation and the Health Professions, 1987
Two types of computer-administered unit quizzes in a systematic pathology course for second-year medical students were compared. Quizzes composed of questions selected on the basis of a student's ability had higher correlations with the final examination than did quizzes composed of questions randomly selected from topic areas. (Author/JAZ)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Difficulty Level
Huntley, Renee M.; Welch, Catherine J. – 1993
Writers of mathematics test items, especially those who write for standardized tests, are often advised to arrange the answer options in logical order, usually ascending or descending numerical order. In this study, 32 mathematics items were selected for inclusion in four experimental pretest units, each consisting of 16 items. Two versions…
Descriptors: Ability, College Entrance Examinations, Comparative Testing, Distractors (Tests)
Melancon, Janet G.; Thompson, Bruce – 1990
Classical measurement theory was used to investigate measurement characteristics of both parts of the Finding Embedded Figures Test (FEFT) when the test was: administered in either a "no guessing" supply format or a multiple-choice selection format; administered to either undergraduate college students or middle school students; and…
Descriptors: Comparative Testing, Construct Validity, Guessing (Tests), Higher Education
Peer reviewedWalstad, William B.; Robson, Denise – Journal of Economic Education, 1997
Applies Item Response Theory methods to data from the national norming of the Test of Economic Literacy to identify test questions with large male-female differences. Regression analysis showed a significant decrease in the magnitude of gender difference, although a difference was still present. (MJP)
Descriptors: Academic Aptitude, Comparative Testing, Economics, Economics Education

Direct link
