Publication Date
| In 2026 | 0 |
| Since 2025 | 10 |
| Since 2022 (last 5 years) | 40 |
| Since 2017 (last 10 years) | 118 |
| Since 2007 (last 20 years) | 211 |
Descriptor
| Multiple Choice Tests | 532 |
| Test Reliability | 532 |
| Test Validity | 302 |
| Test Construction | 238 |
| Test Items | 172 |
| Foreign Countries | 114 |
| Item Analysis | 101 |
| Higher Education | 90 |
| Difficulty Level | 85 |
| Guessing (Tests) | 74 |
| Scoring | 69 |
| More ▼ | |
Source
Author
| Ebel, Robert L. | 10 |
| Frary, Robert B. | 9 |
| Alonzo, Julie | 7 |
| Frisbie, David A. | 6 |
| Irvin, P. Shawn | 6 |
| Lai, Cheng-Fei | 6 |
| Park, Bitnara Jasmine | 6 |
| Tindal, Gerald | 6 |
| Wilcox, Rand R. | 5 |
| Albanese, Mark A. | 4 |
| Biancarosa, Gina | 4 |
| More ▼ | |
Publication Type
Education Level
Audience
| Researchers | 11 |
| Practitioners | 8 |
| Teachers | 5 |
Location
| Indonesia | 17 |
| Turkey | 17 |
| Germany | 8 |
| Iran | 8 |
| Canada | 6 |
| Malaysia | 4 |
| Nigeria | 4 |
| Australia | 3 |
| Florida | 3 |
| Japan | 3 |
| Pakistan | 3 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Xiao, Yang; Han, Jing; Koenig, Kathleen; Xiong, Jianwen; Bao, Lei – Physical Review Physics Education Research, 2018
Assessment instruments composed of two-tier multiple choice (TTMC) items are widely used in science education as an effective method to evaluate students' sophisticated understanding. In practice, however, there are often concerns regarding the common scoring methods of TTMC items, which include pair scoring and individual scoring schemes. The…
Descriptors: Hierarchical Linear Modeling, Item Response Theory, Multiple Choice Tests, Case Studies
Koskey, Kristin L. K.; Makki, Nidaa; Ahmed, Wondimu; Garafolo, Nicholas G.; Visco, Donald P., Jr. – School Science and Mathematics, 2020
Integrating engineering into the K-12 science curriculum continues to be a focus in national reform efforts in science education. Although there is an increasing interest in research in and practice of integrating engineering in K-12 science education, to date only a few studies have focused on the development of an assessment tool to measure…
Descriptors: Middle School Students, Engineering, Design, Science Education
A Feasible Guidance for Ordered Multiple-Choice Items in Students' Hierarchical Understanding Levels
Su, King-Dow – Journal of Baltic Science Education, 2019
This research focuses on students' 5 hierarchical levels of Ordered Multiple-Choice (OMC) items for their extensive conceptualized understanding in the particulate nature of matter (PNM) chemistry. The basic framework for OMC items is to link students' conceptual understanding levels with possible cognitive responses. Developed as the substantial…
Descriptors: Multiple Choice Tests, Science Tests, STEM Education, Test Items
Chin, Huan; Chew, Cheng Meng; Lim, Hooi Lian; Thien, Lei Mee – International Journal of Science and Mathematics Education, 2022
Cognitive Diagnostic Assessment (CDA) is an alternative assessment which can give a clear picture of pupils' learning process and cognitive structures to education stakeholders so that appropriate instructional strategies can be designed to tailored pupils' needs. Coincide with this function, the Ordered Multiple-Choice (OMC) items were…
Descriptors: Mathematics Instruction, Mathematics Tests, Multiple Choice Tests, Diagnostic Tests
Papenberg, Martin; Musch, Jochen – Applied Measurement in Education, 2017
In multiple-choice tests, the quality of distractors may be more important than their number. We therefore examined the joint influence of distractor quality and quantity on test functioning by providing a sample of 5,793 participants with five parallel test sets consisting of items that differed in the number and quality of distractors.…
Descriptors: Multiple Choice Tests, Test Items, Test Validity, Test Reliability
Liampa, Vasiliki; Malandrakis, George N.; Papadopoulou, Penelope; Pnevmatikos, Dimitrios – Research in Science Education, 2019
This study focused on the development and validation of a three-tier multiple-choice diagnostic instrument about the ecological footprint. Each question in the three-tier test comprised by; (a) the content tier, assessing content knowledge; (b) the reason tier, assessing explanatory knowledge; and (c) the confidence tier that differentiates lack…
Descriptors: Test Construction, Test Validity, Multiple Choice Tests, Ecology
Yazdinejad, Anoushe; Zeraatpishe, Mitra – International Journal of Language Testing, 2019
In this study the validity of partial dictation as a measure of overall language proficiency was examined. Two partial dictation tests along with a C-Test, a cloze test, and a reading comprehension test, as criterion measures, were administered to a group of Iranian EFL learners. The coefficients of correlation between partial dictation and…
Descriptors: Test Validity, Verbal Communication, Language Proficiency, Language Tests
Yalaki, Yalçin; Dogan, Nuri; Irez, Serhat; Dogan, Nihal; Çakmakçi, Gültekin; Kara, Basak Erdem – International Journal of Assessment Tools in Education, 2019
Developing scientific literacy for all students is the most often stated purpose of contemporary science education. Nature of science (NOS) is seen as an important component of scientific literacy. There are various perceptions of NOS in the science education community and NOS itself is an ever-changing construct. This makes it challenging to…
Descriptors: Scientific Principles, Scientific Attitudes, Middle School Students, Student Attitudes
Vera Frith; Robert N. Prince – Numeracy, 2018
The National Benchmark Test Project (NBTP) was commissioned by Higher Education South Africa in 2005 to assess the academic proficiency of prospective students. The competencies assessed include quantitative literacy using the NBTP QL test. This instrument is a criterion-referenced multiple-choice test developed collaboratively by South African…
Descriptors: National Competency Tests, Numeracy, Mathematics Tests, Foreign Countries
Perkins, Kyle; Frank, Eva – Online Submission, 2018
This paper presents item-analysis data to illustrate how to identify a set of internally consistent test items that differentiate or discriminate among examinees who are highly proficient and nonproficient on the construct of interest. Suggestions for analyzing the quality of test items are offered as well as a pedagogical approach to augment the…
Descriptors: Item Analysis, Test Items, Test Reliability, Kinetics
Kuo, Bor-Chen; Liao, Chen-Huei; Pai, Kai-Chih; Shih, Shu-Chuan; Li, Cheng-Hsuan; Mok, Magdalena Mo Ching – Educational Psychology, 2020
The current study explores students' collaboration and problem solving (CPS) abilities using a human-to-agent (H-A) computer-based collaborative problem solving assessment. Five CPS assessment units with 76 conversation-based items were constructed using the PISA 2015 CPS framework. In the experiment, 53,855 ninth and tenth graders in Taiwan were…
Descriptors: Computer Assisted Testing, Cooperative Learning, Problem Solving, Item Response Theory
Cromley, Jennifer G.; Dai, Ting; Fechter, Tia; Nelson, Frank E.; Van Boekel, Martin; Du, Yang – Grantee Submission, 2021
Making inferences and reasoning with new scientific information is critical for successful performance in biology coursework. Thus, identifying students who are weak in these skills could allow the early provision of additional support and course placement recommendations to help students develop their reasoning abilities, leading to better…
Descriptors: Science Tests, Multiple Choice Tests, Logical Thinking, Inferences
Haberman, Shelby J.; Liu, Yang; Lee, Yi-Hsuan – ETS Research Report Series, 2019
Distractor analyses are routinely conducted in educational assessments with multiple-choice items. In this research report, we focus on three item response models for distractors: (a) the traditional nominal response (NR) model, (b) a combination of a two-parameter logistic model for item scores and a NR model for selections of incorrect…
Descriptors: Multiple Choice Tests, Scores, Test Reliability, High Stakes Tests
Schuwirth, Lambert W. T.; Van Der Vleuten, Cees P. M. – Journal of Applied Testing Technology, 2019
Programmatic assessment is both a philosophy and a method for assessment. It has been developed in medical education as a response to the limitation of the dominant testing or measurement approaches and to better align with changes in how medical competence was conceptualised. It is based on continual collection of assessment and feedback…
Descriptors: Program Evaluation, Medical Education, Competency Based Education, Feedback (Response)
Lahner, Felicitas-Maria; Lörwald, Andrea Carolin; Bauer, Daniel; Nouns, Zineb Miriam; Krebs, René; Guttormsen, Sissel; Fischer, Martin R.; Huwendiek, Sören – Advances in Health Sciences Education, 2018
Multiple true-false (MTF) items are a widely used supplement to the commonly used single-best answer (Type A) multiple choice format. However, an optimal scoring algorithm for MTF items has not yet been established, as existing studies yielded conflicting results. Therefore, this study analyzes two questions: What is the optimal scoring algorithm…
Descriptors: Scoring Formulas, Scoring Rubrics, Objective Tests, Multiple Choice Tests

Peer reviewed
Direct link
