Publication Date
| In 2026 | 0 |
| Since 2025 | 4 |
| Since 2022 (last 5 years) | 28 |
| Since 2017 (last 10 years) | 68 |
| Since 2007 (last 20 years) | 142 |
Descriptor
| Testing | 442 |
| Test Validity | 334 |
| Test Reliability | 152 |
| Test Construction | 100 |
| Language Tests | 92 |
| Foreign Countries | 85 |
| English (Second Language) | 64 |
| Second Language Learning | 62 |
| Higher Education | 60 |
| Scores | 58 |
| Validity | 53 |
| More ▼ | |
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 13 |
| Researchers | 11 |
| Teachers | 4 |
| Administrators | 1 |
| Counselors | 1 |
Location
| Canada | 9 |
| China | 7 |
| Australia | 6 |
| California | 5 |
| United Kingdom | 5 |
| Illinois | 4 |
| Iran | 4 |
| Pennsylvania | 4 |
| Turkey | 4 |
| United Kingdom (England) | 4 |
| Brazil | 3 |
| More ▼ | |
Laws, Policies, & Programs
| No Child Left Behind Act 2001 | 2 |
| Bilingual Education Act 1968 | 1 |
| Elementary and Secondary… | 1 |
| Elementary and Secondary… | 1 |
| Elementary and Secondary… | 1 |
| Race to the Top | 1 |
| Rehabilitation Act 1973… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Sherwin E. Balbuena – Online Submission, 2024
This study introduces a new chi-square test statistic for testing the equality of response frequencies among distracters in multiple-choice tests. The formula uses the information from the number of correct answers and wrong answers, which becomes the basis of calculating the expected values of response frequencies per distracter. The method was…
Descriptors: Multiple Choice Tests, Statistics, Test Validity, Testing
Gökhan Iskifoglu – Turkish Online Journal of Educational Technology - TOJET, 2024
This research paper investigated the importance of conducting measurement invariance analysis in developing measurement tools for assessing differences between and among study variables. Most of the studies, which tended to develop an inventory to assess the existence of an attitude, behavior, belief, IQ, or an intuition in a person's…
Descriptors: Testing, Testing Problems, Error of Measurement, Attitude Measures
Amanda A. Wolkowitz; Russell Smith – Practical Assessment, Research & Evaluation, 2024
A decision consistency (DC) index is an estimate of the consistency of a classification decision on an exam. More specifically, DC estimates the percentage of examinees that would have the same classification decision on an exam if they were to retake the same or a parallel form of the exam again without memory of taking the exam the first time.…
Descriptors: Testing, Test Reliability, Replication (Evaluation), Decision Making
James Soland – Journal of Research on Educational Effectiveness, 2024
When randomized control trials are not possible, quasi-experimental methods often represent the gold standard. One quasi-experimental method is difference-in-difference (DiD), which compares changes in outcomes before and after treatment across groups to estimate a causal effect. DiD researchers often use fairly exhaustive robustness checks to…
Descriptors: Item Response Theory, Testing, Test Validity, Intervention
Maciej Koscielniak; Jolanta Enko; Agata Gasiorowska – Journal of Academic Ethics, 2024
Examination dishonesty is a global problem that became particularly critical after the outbreak of the COVID-19 pandemic and the shift to remote learning. Academic research has often examined this phenomenon as only one aspect of a broader concept of academic dishonesty and as a one-dimensional construct. This article builds on existing knowledge…
Descriptors: Foreign Countries, Students, Ethics, Cheating
Peabody, Michael R.; Muckle, Timothy J.; Meng, Yu – Educational Measurement: Issues and Practice, 2023
The subjective aspect of standard-setting is often criticized, yet data-driven standard-setting methods are rarely applied. Therefore, we applied a mixture Rasch model approach to setting performance standards across several testing programs of various sizes and compared the results to existing passing standards derived from traditional…
Descriptors: Item Response Theory, Standard Setting, Testing, Sampling
Ian Phil Canlas; Joyce Molino-Magtolis – Journal of Biological Education, 2024
The use of drawing as an assessment tool to reveal students' conceptions in biology specifically on human organs and organ systems is not new, however, there is a deficit in the literature that attempted to explore and reflect on its usefulness and relevance specifically, in eliciting students' preconceptions related thereto. Making use of a…
Descriptors: Foreign Countries, Preservice Teacher Education, Preservice Teachers, Biology
Jeff Allen; Jay Thomas; Stacy Dreyer; Scott Johanningmeier; Dana Murano; Ty Cruce; Xin Li; Edgar Sanchez – ACT Education Corp., 2025
This report describes the process of developing and validating the enhanced ACT. The report describes the changes made to the test content and the processes by which these design decisions were implemented. The authors describe how they shared the overall scope of the enhancements, including the initial blueprints, with external expert panels,…
Descriptors: College Entrance Examinations, Testing, Change, Test Construction
Militsa G. Ivanova; Michalis P. Michaelides – Practical Assessment, Research & Evaluation, 2023
Research on methods for measuring examinee engagement with constructed-response items is limited. The present study used data from the PISA 2018 Reading domain to construct and compare indicators of test-taking effort on constructed-response items: response time, number of actions, the union (combining effortless responses detected by either…
Descriptors: Achievement Tests, Foreign Countries, International Assessment, Secondary School Students
Dongmei Li; Shalini Kapoor; Ann Arthur; Chi-Yu Huang; YoungWoo Cho; Chen Qiu; Hongling Wang – ACT Education Corp., 2025
Starting in April 2025, ACT will introduce enhanced forms of the ACT® test for national online testing, with a full rollout to all paper and online test takers in national, state and district, and international test administrations by Spring 2026. ACT introduced major updates by changing the test lengths and testing times, providing more time per…
Descriptors: College Entrance Examinations, Testing, Change, Scoring
Jeff Allen; Ty Cruce – ACT Education Corp., 2025
This report summarizes some of the evidence supporting interpretations of scores from the enhanced ACT, focusing on reliability, concurrent validity, predictive validity, and score comparability. The authors argue that the evidence presented in this report supports the interpretation of scores from the enhanced ACT as measures of high school…
Descriptors: College Entrance Examinations, Testing, Change, Scores
Vahed, Anisa; Walters, Matthys Michielse; Ross, Ashley Hilton Adrian – Education Inquiry, 2023
Despite the expanding literature in the last three decades on modes of implementation and the various forms of formal and informal assessments, there is limited evidence of academics' knowledge and understanding of continuous assessment practice. Using a mixed methods sequential explanatory research design, this paper aimed to investigate…
Descriptors: Assessment Literacy, Student Evaluation, Formative Evaluation, Summative Evaluation
Shun-Fu Hu; Amery D. Wu; Jake Stone – Journal of Educational Measurement, 2025
Scoring high-dimensional assessments (e.g., > 15 traits) can be a challenging task. This paper introduces the multilabel neural network (MNN) as a scoring method for high-dimensional assessments. Additionally, it demonstrates how MNN can score the same test responses to maximize different performance metrics, such as accuracy, recall, or…
Descriptors: Tests, Testing, Scores, Test Construction
Lies Sercu – Discover Education, 2023
Health Literacy (HL) refers to personal competencies for accessing, understanding, appraising, and applying health-relevant information in order to make sound decisions in everyday life that promote the prevention of bad and maintenance of good health. Schools are seen as promising intervention settings for promoting these skills in adolescents.…
Descriptors: Factor Analysis, Construct Validity, Health, Knowledge Level
Meagan Karvonen; Russell Swinburne Romine; Amy K. Clark – Practical Assessment, Research & Evaluation, 2024
This paper describes methods and findings from student cognitive labs, teacher cognitive labs, and test administration observations as evidence evaluated in a validity argument for a computer-based alternate assessment for students with significant cognitive disabilities. Validity of score interpretations and uses for alternate assessments based…
Descriptors: Students with Disabilities, Intellectual Disability, Severe Disabilities, Student Evaluation

Peer reviewed
Direct link
