Publication Date
In 2025 | 1 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 9 |
Since 2016 (last 10 years) | 18 |
Since 2006 (last 20 years) | 35 |
Descriptor
Comparative Analysis | 53 |
Difficulty Level | 53 |
Test Format | 53 |
Test Items | 37 |
Foreign Countries | 18 |
Item Response Theory | 17 |
Multiple Choice Tests | 17 |
Item Analysis | 15 |
Computer Assisted Testing | 14 |
Language Tests | 13 |
Scores | 11 |
More ▼ |
Source
Author
Allen, Nancy L. | 2 |
DeBoer, George E. | 2 |
Hardcastle, Joseph | 2 |
Herrmann-Abell, Cari F. | 2 |
Kim, Sooyeon | 2 |
Wainer, Howard | 2 |
Alpayar, Cagla | 1 |
Apino, Ezi | 1 |
Babiar, Tasha Calvert | 1 |
Baghaei, Purya | 1 |
Batty, Aaron Olaf | 1 |
More ▼ |
Publication Type
Reports - Research | 44 |
Journal Articles | 34 |
Speeches/Meeting Papers | 15 |
Reports - Evaluative | 5 |
Tests/Questionnaires | 3 |
Dissertations/Theses -… | 2 |
Collected Works - Proceedings | 1 |
Information Analyses | 1 |
Education Level
Higher Education | 11 |
Postsecondary Education | 11 |
Secondary Education | 5 |
High Schools | 3 |
Elementary Education | 2 |
Elementary Secondary Education | 1 |
Grade 12 | 1 |
Grade 8 | 1 |
Intermediate Grades | 1 |
Middle Schools | 1 |
Audience
Location
Indonesia | 2 |
Japan | 2 |
Spain | 2 |
United Kingdom (England) | 2 |
Canada | 1 |
China | 1 |
Europe | 1 |
Germany | 1 |
India | 1 |
Iran | 1 |
Macau | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Harrison, Scott; Kroehne, Ulf; Goldhammer, Frank; Lüdtke, Oliver; Robitzsch, Alexander – Large-scale Assessments in Education, 2023
Background: Mode effects, the variations in item and scale properties attributed to the mode of test administration (paper vs. computer), have stimulated research around test equivalence and trend estimation in PISA. The PISA assessment framework provides the backbone to the interpretation of the results of the PISA test scores. However, an…
Descriptors: Scoring, Test Items, Difficulty Level, Foreign Countries
Herrmann-Abell, Cari F.; Hardcastle, Joseph; DeBoer, George E. – Grantee Submission, 2022
As implementation of the "Next Generation Science Standards" moves forward, there is a need for new assessments that can measure students' integrated three-dimensional science learning. The National Research Council has suggested that these assessments be multicomponent tasks that utilize a combination of item formats including…
Descriptors: Multiple Choice Tests, Conditioning, Test Items, Item Response Theory
Kárász, Judit T.; Széll, Krisztián; Takács, Szabolcs – Quality Assurance in Education: An International Perspective, 2023
Purpose: Based on the general formula, which depends on the length and difficulty of the test, the number of respondents and the number of ability levels, this study aims to provide a closed formula for the adaptive tests with medium difficulty (probability of solution is p = 1/2) to determine the accuracy of the parameters for each item and in…
Descriptors: Test Length, Probability, Comparative Analysis, Difficulty Level
Musa Adekunle Ayanwale – Discover Education, 2023
Examination scores obtained by students from the West African Examinations Council (WAEC), and National Business and Technical Examinations Board (NABTEB) may not be directly comparable due to differences in examination administration, item characteristics of the subject in question, and student abilities. For more accurate comparisons, scores…
Descriptors: Equated Scores, Mathematics Tests, Test Items, Test Format
Benton, Tom – Research Matters, 2021
Computer adaptive testing is intended to make assessment more reliable by tailoring the difficulty of the questions a student has to answer to their level of ability. Most commonly, this benefit is used to justify the length of tests being shortened whilst retaining the reliability of a longer, non-adaptive test. Improvements due to adaptive…
Descriptors: Risk, Item Response Theory, Computer Assisted Testing, Difficulty Level
Yangqiuting Li; Chandralekha Singh – Physical Review Physics Education Research, 2025
Research-based multiple-choice questions implemented in class with peer instruction have been shown to be an effective tool for improving students' engagement and learning outcomes. Moreover, multiple-choice questions that are carefully sequenced to build on each other can be particularly helpful for students to develop a systematic understanding…
Descriptors: Physics, Science Instruction, Science Tests, Multiple Choice Tests
Herrmann-Abell, Cari F.; Hardcastle, Joseph; DeBoer, George E. – Grantee Submission, 2019
The "Next Generation Science Standards" calls for new assessments that measure students' integrated three-dimensional science learning. The National Research Council has suggested that these assessments utilize a combination of item formats including constructed-response and multiple-choice. In this study, students were randomly assigned…
Descriptors: Science Tests, Multiple Choice Tests, Test Format, Test Items
Yao, Don – English Language Teaching, 2020
Computer-based test (CBT) and paper-based test (PBT) are two test modes to the test takers that have been widely adopted in the field of language testing or assessment over the last few decades. Due to the rapid development of science and technology, it is a trend for universities and educational institutions striving rather hard to deliver the…
Descriptors: Language Tests, Computer Assisted Testing, Test Format, Comparative Analysis
Kim, Ahyoung Alicia; Tywoniw, Rurik L.; Chapman, Mark – Language Assessment Quarterly, 2022
Technology-enhanced items (TEIs) are innovative, computer-delivered test items that allow test takers to better interact with the test environment compared to traditional multiple-choice items (MCIs). The interactive nature of TEIs offer improved construct coverage compared with MCIs but little research exists regarding students' performance on…
Descriptors: Language Tests, Test Items, Computer Assisted Testing, English (Second Language)
Liao, Linyu – English Language Teaching, 2020
As a high-stakes standardized test, IELTS is expected to have comparable forms of test papers so that test takers from different test administration on different dates receive comparable test scores. Therefore, this study examined the text difficulty and task characteristics of four parallel academic IELTS reading tests to reveal to what extent…
Descriptors: Second Language Learning, English (Second Language), Language Tests, High Stakes Tests
Gu, Lin; Ling, Guangming; Liu, Ou Lydia; Yang, Zhitong; Li, Guirong; Kardanova, Elena; Loyalka, Prashant – Assessment & Evaluation in Higher Education, 2021
We examine the effects of computer-based versus paper-based assessment of critical thinking skills, adapted from English (in the U.S.) to Chinese. Using data collected based on a random assignment between the two modes in multiple Chinese colleges, we investigate mode effects from multiple perspectives: mean scores, measurement precision, item…
Descriptors: Critical Thinking, Tests, Test Format, Computer Assisted Testing
Yeager, Rebecca; Meyer, Zachary – International Journal of Listening, 2022
This study investigates the effects of adding stem preview to an English for Academic Purposes (EAP) multiple-choice listening assessment. In stem preview, listeners may view the item stems, but not response options, before listening. Previous research indicates that adding preview to an exam typically decreases difficulty, but raises concerns…
Descriptors: English for Academic Purposes, Second Language Learning, Second Language Instruction, Teaching Methods
Lahner, Felicitas-Maria; Lörwald, Andrea Carolin; Bauer, Daniel; Nouns, Zineb Miriam; Krebs, René; Guttormsen, Sissel; Fischer, Martin R.; Huwendiek, Sören – Advances in Health Sciences Education, 2018
Multiple true-false (MTF) items are a widely used supplement to the commonly used single-best answer (Type A) multiple choice format. However, an optimal scoring algorithm for MTF items has not yet been established, as existing studies yielded conflicting results. Therefore, this study analyzes two questions: What is the optimal scoring algorithm…
Descriptors: Scoring Formulas, Scoring Rubrics, Objective Tests, Multiple Choice Tests
Elfiondri; Kasim, Usman; Mustafa, Faisal; Putra, Tomi Mandala – TESOL International Journal, 2020
Studies have shown that reading comprehension is the most difficult section of the Paper-Based Test (PBT) TOEFL. Therefore, this research aimed to identify which sub-skill in reading comprehension poses the greatest challenges for the students and how this sub-skill correlates to other reading comprehension sub-skills. To achieve this purpose,…
Descriptors: Reading Comprehension, Second Language Learning, Language Tests, English (Second Language)
Madya, Suwarsih; Retnawati, Heri; Purnawan, Ari; Putro, Nur Hidayanto Pancoro Setyo; Apino, Ezi – TEFLIN Journal: A publication on the teaching and learning of English, 2019
This explorative-descriptive study set out to examine the equivalence among Test of English Proficiency (TOEP) forms, developed by the Indonesian Testing Service Centre (ITSC) and co-founded by The Association for The Teaching of English as a Foreign Language in Indonesia (TEFLIN) and The Association of Psychology in Indonesia. Using a…
Descriptors: Language Tests, Language Proficiency, English (Second Language), Second Language Learning