Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 2 |
| Since 2007 (last 20 years) | 14 |
Descriptor
Source
Author
Publication Type
| Reports - Evaluative | 56 |
| Journal Articles | 32 |
| Speeches/Meeting Papers | 9 |
| Information Analyses | 3 |
| Reports - Research | 2 |
| Books | 1 |
| Guides - Classroom - Teacher | 1 |
| Guides - General | 1 |
| Guides - Non-Classroom | 1 |
| Opinion Papers | 1 |
Education Level
| Higher Education | 5 |
| High Schools | 4 |
| Secondary Education | 4 |
| Elementary Education | 3 |
| Elementary Secondary Education | 3 |
| Middle Schools | 2 |
| Postsecondary Education | 2 |
| Adult Education | 1 |
| Grade 12 | 1 |
| Grade 5 | 1 |
| Grade 8 | 1 |
| More ▼ | |
Audience
| Practitioners | 3 |
| Teachers | 2 |
| Researchers | 1 |
Location
| Netherlands | 3 |
| Australia | 2 |
| California | 1 |
| Japan | 1 |
| Malaysia | 1 |
| New Jersey | 1 |
| Singapore | 1 |
| United Arab Emirates | 1 |
| United States | 1 |
| Virginia | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Toker, Deniz – TESL-EJ, 2019
The central purpose of this paper is to examine validity problems arising from the multiple-choice items and technical passages in the Test of English as a Foreign Language Internet-based Test (TOEFL iBT) reading section, primarily concentrating on construct-irrelevant variance (Messick, 1989). My personal TOEFL iBT experience, along with my…
Descriptors: English (Second Language), Language Tests, Second Language Learning, Computer Assisted Testing
Wise, Steven L. – Educational Measurement: Issues and Practice, 2017
The rise of computer-based testing has brought with it the capability to measure more aspects of a test event than simply the answers selected or constructed by the test taker. One behavior that has drawn much research interest is the time test takers spend responding to individual multiple-choice items. In particular, very short response…
Descriptors: Guessing (Tests), Multiple Choice Tests, Test Items, Reaction Time
Coniam, David – Frontiers of Education in China, 2014
This article examines the issue of the quality of teacher-produced tests, limiting itself in the current context to objective, multiple-choice tests. The article investigates a short, two-part 20-item English language test. After a brief overview of the key test qualities of reliability and validity, the article examines the two subtests in terms…
Descriptors: Teacher Made Tests, English (Second Language), Language Tests, Multiple Choice Tests
Stewart, Jeffrey – Language Assessment Quarterly, 2014
Validated under a Rasch framework (Beglar, 2010), the Vocabulary Size Test (VST) (Nation & Beglar, 2007) is an increasingly popular measure of decontextualized written receptive vocabulary size in the field of second language acquisition. However, although the validation indicates that the test has high internal reliability, still unaddressed…
Descriptors: Multiple Choice Tests, Vocabulary, Language Tests, Receptive Language
Williamson, Kathryn E.; Willoughby, Shannon; Prather, Edward E. – Astronomy Education Review, 2013
We introduce the Newtonian Gravity Concept Inventory (NGCI), a 26-item multiple-choice instrument to assess introductory general education college astronomy ("Astro 101") student understanding of Newtonian gravity. This paper describes the development of the NGCI through four phases: Planning, Construction, Quantitative Analysis, and…
Descriptors: Science Instruction, Scientific Concepts, Astronomy, College Science
Coombe, Christine; Davidson, Peter – Language Testing, 2014
The Common Educational Proficiency Assessment (CEPA) is a large-scale, high-stakes, English language proficiency/placement test administered in the United Arab Emirates to Emirati nationals in their final year of secondary education or Grade 12. The purpose of the CEPA is to place students into English classes at the appropriate government…
Descriptors: Language Tests, High Stakes Tests, English (Second Language), Second Language Learning
Wan, Lei; Henly, George A. – Applied Measurement in Education, 2012
Many innovative item formats have been proposed over the past decade, but little empirical research has been conducted on their measurement properties. This study examines the reliability, efficiency, and construct validity of two innovative item formats--the figural response (FR) and constructed response (CR) formats used in a K-12 computerized…
Descriptors: Test Items, Test Format, Computer Assisted Testing, Measurement
Hamzah, Mohd Sahandri Gani; Abdullah, Saifuddin Kumar – Online Submission, 2011
The evaluation of learning is a systematic process involving testing, measuring and evaluation. In the testing step, a teacher needs to choose the best instrument that can test the minds of students. Testing will produce scores or marks with many variations either in homogeneous or heterogeneous forms that will be used to categorize the scores…
Descriptors: Test Items, Item Analysis, Difficulty Level, Testing
Malau-Aduli, Bunmi S.; Zimitat, Craig – Assessment & Evaluation in Higher Education, 2012
The aim of this study was to assess the effect of the introduction of peer review processes on the quality of multiple-choice examinations in the first three years of an Australian medical course. The impact of the peer review process and overall quality assurance (QA) processes were evaluated by comparing the examination data generated in earlier…
Descriptors: Foreign Countries, Peer Evaluation, Multiple Choice Tests, Test Construction
Gekara, Victor Oyaro; Bloor, Michael; Sampson, Helen – Journal of Vocational Education and Training, 2011
Vocational education and training (VET) concerns the cultivation and development of specific skills and competencies, in addition to broad underpinning knowledge relating to paid employment. VET assessment is, therefore, designed to determine the extent to which a trainee has effectively acquired the knowledge, skills, and competencies required by…
Descriptors: Marine Education, Occupational Safety and Health, Computer Assisted Testing, Vocational Education
Reshetar, Rosemary; Melican, Gerald J. – College Board, 2010
This paper discusses issues related to the design and psychometric work for mixed-format tests --tests containing both multiple-choice (MC) and constructed-response (CR) items. The issues of validity, fairness, reliability and score consistency can be addressed but for mixed-format tests there are many decisions to be made and no examination or…
Descriptors: Psychometrics, Test Construction, Multiple Choice Tests, Test Items
Coleman, Chris; Lindstrom, Jennifer; Nelson, Jason; Lindstrom, William; Gregg, K. Noel – Journal of Learning Disabilities, 2010
The comprehension section of the "Nelson-Denny Reading Test" (NDRT) is widely used to assess the reading comprehension skills of adolescents and adults in the United States. In this study, the authors explored the content validity of the NDRT Comprehension Test (Forms G and H) by asking university students (with and without at-risk…
Descriptors: Reading Comprehension, Reading Difficulties, Reading Tests, Content Validity
Trieber, J. Marshall – Training and Development Journal, 1980
Aims to help instructors make more valid test questions, particularly multiple-choice ones. Emphasis is placed on multiple-choice questions to show the wealth of opportunities they offer for testing because of their uses, objectivity, and ease of grading. Discusses test scheduling, construction, and evaluation and follow-up. (CT)
Descriptors: Multiple Choice Tests, Test Construction, Test Reliability, Test Validity
Peer reviewedTuckman, Bruce W. – NASSP Bulletin, 1993
Essay tests are easily constructed, relatively valid assessments of higher cognitive processes but are harder to score reliably. Teachers using essay tests are advised to follow clearly designed objectives, construct all-inclusive, pilot-tested questions, develop a checklist of specific scoring points and a model answer for each question, and use…
Descriptors: Essay Tests, Multiple Choice Tests, Scoring, Secondary Education
Marrelli, Anne F. – Performance and Instruction, 1995
Discusses the advantages of using multiple choice questions, highlighting the flexibility of using different variations of questions. Item writing guidelines include information on content, sensitivity, difficulty, irrelevant sources of difficulty, order, misleads, avoidance of clues, and exercises in the application of guidelines. (JKP)
Descriptors: Distractors (Tests), Guidelines, Multiple Choice Tests, Questioning Techniques

Direct link
