Publication Date
| In 2026 | 0 |
| Since 2025 | 8 |
| Since 2022 (last 5 years) | 61 |
| Since 2017 (last 10 years) | 125 |
| Since 2007 (last 20 years) | 219 |
Descriptor
| Test Construction | 813 |
| Test Format | 813 |
| Test Items | 363 |
| Test Validity | 180 |
| Higher Education | 177 |
| Computer Assisted Testing | 146 |
| Multiple Choice Tests | 139 |
| Test Reliability | 135 |
| Foreign Countries | 130 |
| Elementary Secondary Education | 108 |
| Language Tests | 87 |
| More ▼ | |
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 78 |
| Teachers | 57 |
| Researchers | 38 |
| Administrators | 16 |
| Students | 6 |
| Policymakers | 5 |
| Media Staff | 1 |
| Parents | 1 |
Location
| Turkey | 12 |
| Canada | 10 |
| Japan | 10 |
| United States | 9 |
| United Kingdom | 8 |
| Germany | 7 |
| Australia | 6 |
| Israel | 6 |
| California | 5 |
| China | 5 |
| Florida | 5 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 3 |
| Improving Americas Schools… | 1 |
| Individuals with Disabilities… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
National Assessment Governing Board, 2014
Since 1973, the National Assessment of Educational Progress (NAEP) has gathered information about student achievement in mathematics. Results of these periodic assessments, produced in print and web-based formats, provide valuable information to a wide variety of audiences. They inform citizens about the nature of students' comprehension of the…
Descriptors: National Competency Tests, Mathematics Achievement, Mathematics Skills, Grade 4
Kitazawa, Takeshi – International Association for Development of the Information Society, 2013
This study analyzed effective test formats that utilized tablets for tests in university information basic subjects in blended learning environments. Specifically, three types of test were created: (1) multiple-choice, (2) fill-in-the-blank, and (3) a mixture of multiple-choice and fill-in-the-blank. An analysis focusing on university students'…
Descriptors: Test Format, Test Construction, Handheld Devices, Computers
Williams, Marian E.; Sando, Lara; Soles, Tamara Glen – Journal of Psychoeducational Assessment, 2014
Cognitive assessment of young children contributes to high-stakes decisions because results are often used to determine eligibility for early intervention and special education. Previous reviews of cognitive measures for young children highlighted concerns regarding adequacy of standardization samples, steep item gradients, and insufficient floors…
Descriptors: Intelligence Tests, Decision Making, High Stakes Tests, Eligibility
Lee, Eunjung; Lee, Won-Chan; Brennan, Robert L. – College Board, 2012
In almost all high-stakes testing programs, test equating is necessary to ensure that test scores across multiple test administrations are equivalent and can be used interchangeably. Test equating becomes even more challenging in mixed-format tests, such as Advanced Placement Program® (AP®) Exams, that contain both multiple-choice and constructed…
Descriptors: Test Construction, Test Interpretation, Test Norms, Test Reliability
McLean, Stuart; Kramer, Brandon; Beglar, David – Language Teaching Research, 2015
An important gap in the field of second language vocabulary assessment concerns the lack of validated tests measuring aural vocabulary knowledge. The primary purpose of this study is to introduce and provide preliminary validity evidence for the Listening Vocabulary Levels Test (LVLT), which has been designed as a diagnostic tool to measure…
Descriptors: Test Construction, Test Validity, English (Second Language), Second Language Learning
Albano, Anthony D. – Journal of Educational Measurement, 2013
In many testing programs it is assumed that the context or position in which an item is administered does not have a differential effect on examinee responses to the item. Violations of this assumption may bias item response theory estimates of item and person parameters. This study examines the potentially biasing effects of item position. A…
Descriptors: Test Items, Item Response Theory, Test Format, Questioning Techniques
Plassmann, Sibylle; Zeidler, Beate – Language Learning in Higher Education, 2014
Language testing means taking decisions: about the test taker's results, but also about the test construct and the measures taken in order to ensure quality. This article takes the German test "telc Deutsch C1 Hochschule" as an example to illustrate this decision-making process in an academic context. The test is used for university…
Descriptors: Language Tests, Test Wiseness, Test Construction, Decision Making
Post, Gerald V.; Hargis, Jace – Decision Sciences Journal of Innovative Education, 2012
Online education and computer-assisted instruction (CAI) have existed for years, but few general tools exist to help instructors create and evaluate lessons. Are these tools sufficient? Specifically, what elements do instructors want to see in online testing tools? This study asked instructors from various disciplines to identify and evaluate the…
Descriptors: Computer Assisted Testing, Computer Software, Test Construction, Design Preferences
Mullis, Ina V. S., Ed.; Martin, Michael O., Ed. – International Association for the Evaluation of Educational Achievement, 2014
It is critical for countries to ensure that capable secondary school students receive further preparation in advanced mathematics and science, so that they are ready to enter challenging university-level studies that prepare them for careers in science, technology, engineering, and mathematics (STEM) fields. This group of students will become the…
Descriptors: Mathematics Tests, Science Tests, Educational Assessment, Secondary School Students
Jeanne Carol Samuel – ProQuest LLC, 2012
Large numbers of students attending community college lack essential college success skills (motivation and study strategies). Many of these students do not complete their degree programs. Identifying learning and teaching methods that promote the development of lifelong learning skills in addition to content acquisition is essential. This…
Descriptors: Test Construction, Student Motivation, Learning Strategies, Community College Students
Breakstone, Joel – Theory and Research in Social Education, 2014
This article considers the design process for new formative history assessments. Over the course of 3 years, my colleagues from the Stanford History Education Group and I designed, piloted, and revised dozens of "History Assessments of Thinking" (HATs). As we created HATs, we sought to gather information about their cognitive validity,…
Descriptors: History Instruction, Formative Evaluation, Tests, Correlation
Becker, Kirk A.; Bergstrom, Betty A. – Practical Assessment, Research & Evaluation, 2013
The need for increased exam security, improved test formats, more flexible scheduling, better measurement, and more efficient administrative processes has caused testing agencies to consider converting the administration of their exams from paper-and-pencil to computer-based testing (CBT). Many decisions must be made in order to provide an optimal…
Descriptors: Testing, Models, Testing Programs, Program Administration
DeStefano, Lizanne; Johnson, Jeremiah – American Institutes for Research, 2013
This paper describes one of the first efforts by the National Assessment of Educational Progress (NAEP) to improve measurement at the lower end of the distribution, including measurement for students with disabilities (SD) and English language learners (ELLs). One way to improve measurement at the lower end is to introduce one or more…
Descriptors: National Competency Tests, Measures (Individuals), Disabilities, English Language Learners
van der Linden, Wim J.; Diao, Qi – Journal of Educational Measurement, 2011
In automated test assembly (ATA), the methodology of mixed-integer programming is used to select test items from an item bank to meet the specifications for a desired test form and optimize its measurement accuracy. The same methodology can be used to automate the formatting of the set of selected items into the actual test form. Three different…
Descriptors: Test Items, Test Format, Test Construction, Item Banks
O'Reilly, Tenaha; Sabatini, John – ETS Research Report Series, 2013
This paper represents the third installment of the Reading for Understanding (RfU) assessment framework. This paper builds upon the two prior installments (Sabatini & O'Reilly, 2013; Sabatini, O'Reilly, & Deane, 2013) by discussing the role of performance moderators in the test design and how scenario-based assessment can be used as a tool…
Descriptors: Reading Comprehension, Reading Tests, Test Construction, Student Characteristics

Peer reviewed
Direct link
