Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 3 |
Since 2016 (last 10 years) | 9 |
Since 2006 (last 20 years) | 12 |
Descriptor
Source
Author
Publication Type
Journal Articles | 13 |
Reports - Research | 11 |
Reports - Evaluative | 5 |
Reports - Descriptive | 2 |
Speeches/Meeting Papers | 2 |
Collected Works - General | 1 |
Guides - General | 1 |
Numerical/Quantitative Data | 1 |
Tests/Questionnaires | 1 |
Education Level
Grade 8 | 7 |
Elementary Secondary Education | 6 |
Secondary Education | 6 |
Elementary Education | 5 |
Junior High Schools | 5 |
Middle Schools | 5 |
Grade 4 | 3 |
Intermediate Grades | 2 |
Audience
Location
Ireland | 2 |
United States | 2 |
Hong Kong | 1 |
Netherlands | 1 |
Spain | 1 |
Sweden | 1 |
Taiwan | 1 |
Turkey | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Trends in International… | 18 |
Program for International… | 1 |
What Works Clearinghouse Rating
Gustafsson, Martin; Barakat, Bilal Fouad – Comparative Education Review, 2023
International assessments inform education policy debates, yet little is known about their floor effects: To what extent do they fail to differentiate between the lowest performers, and what are the implications of this? TIMSS, SACMEQ, and LLECE data are analyzed to answer this question. In TIMSS, floor effects have been reduced through the…
Descriptors: Achievement Tests, Elementary Secondary Education, International Assessment, Foreign Countries
Lawrence T. DeCarlo – Educational and Psychological Measurement, 2024
A psychological framework for different types of items commonly used with mixed-format exams is proposed. A choice model based on signal detection theory (SDT) is used for multiple-choice (MC) items, whereas an item response theory (IRT) model is used for open-ended (OE) items. The SDT and IRT models are shown to share a common conceptualization…
Descriptors: Test Format, Multiple Choice Tests, Item Response Theory, Models
Ilhan, Mustafa; Öztürk, Nagihan Boztunç; Sahin, Melek Gülsah – Participatory Educational Research, 2020
In this research, the effect of an item's type and cognitive level on its difficulty index was investigated. The data source of the study consisted of the responses of the 12535 students in the Turkey sample (6079 and 6456 students from eighth and fourth grade respectively) of TIMSS 2015. The responses were a total of 215 items at the eighth-grade…
Descriptors: Test Items, Difficulty Level, Cognitive Processes, Responses
Liou, Pey-Yan; Bulut, Okan – Research in Science Education, 2020
The purpose of this study was to examine eighth-grade students' science performance in terms of two test design components, item format, and cognitive domain. The portion of Taiwanese data came from the 2011 administration of the Trends in International Mathematics and Science Study (TIMSS), one of the major international large-scale assessments…
Descriptors: Foreign Countries, Middle School Students, Grade 8, Science Achievement
Hamhuis, Eva; Glas, Cees; Meelissen, Martina – British Journal of Educational Technology, 2020
Over the last two decades, the educational use of digital devices, including digital assessments, has become a regular feature of teaching in primary education in the Netherlands. However, researchers have not reached a consensus about the so-called "mode effect," which refers to the possible impact of using computer-based tests (CBT)…
Descriptors: Handheld Devices, Elementary School Students, Grade 4, Foreign Countries
Wang, Yan; Kim, Eun Sook; Dedrick, Robert F.; Ferron, John M.; Tan, Tony – Educational and Psychological Measurement, 2018
Wording effects associated with positively and negatively worded items have been found in many scales. Such effects may threaten construct validity and introduce systematic bias in the interpretation of results. A variety of models have been applied to address wording effects, such as the correlated uniqueness model and the correlated traits and…
Descriptors: Test Items, Test Format, Correlation, Construct Validity
Mullis, Ina V. S., Ed.; Martin, Michael O., Ed.; von Davier, Matthias, Ed. – International Association for the Evaluation of Educational Achievement, 2021
TIMSS (Trends in International Mathematics and Science Study) is a long-standing international assessment of mathematics and science at the fourth and eighth grades that has been collecting trend data every four years since 1995. About 70 countries use TIMSS trend data for monitoring the effectiveness of their education systems in a global…
Descriptors: Achievement Tests, International Assessment, Science Achievement, Mathematics Achievement
Martin, Michael O., Ed.; von Davier, Matthias, Ed.; Mullis, Ina V. S., Ed. – International Association for the Evaluation of Educational Achievement, 2020
The chapters in this online volume comprise the TIMSS & PIRLS International Study Center's technical report of the methods and procedures used to develop, implement, and report the results of TIMSS 2019. There were various technical challenges because TIMSS 2019 was the initial phase of the transition to eTIMSS, with approximately half the…
Descriptors: Foreign Countries, Elementary Secondary Education, Achievement Tests, International Assessment
Wang, Wen-Chung; Chen, Hui-Fang; Jin, Kuan-Yu – Educational and Psychological Measurement, 2015
Many scales contain both positively and negatively worded items. Reverse recoding of negatively worded items might not be enough for them to function as positively worded items do. In this study, we commented on the drawbacks of existing approaches to wording effect in mixed-format scales and used bi-factor item response theory (IRT) models to…
Descriptors: Item Response Theory, Test Format, Language Usage, Test Items
Öztürk-Gübes, Nese; Kelecioglu, Hülya – Educational Sciences: Theory and Practice, 2016
The purpose of this study was to examine the impact of dimensionality, common-item set format, and different scale linking methods on preserving equity property with mixed-format test equating. Item response theory (IRT) true-score equating (TSE) and IRT observed-score equating (OSE) methods were used under common-item nonequivalent groups design.…
Descriptors: Test Format, Item Response Theory, True Scores, Equated Scores
Mullis, Ina V. S., Ed.; Martin, Michael O., Ed. – International Association for the Evaluation of Educational Achievement, 2014
It is critical for countries to ensure that capable secondary school students receive further preparation in advanced mathematics and science, so that they are ready to enter challenging university-level studies that prepare them for careers in science, technology, engineering, and mathematics (STEM) fields. This group of students will become the…
Descriptors: Mathematics Tests, Science Tests, Educational Assessment, Secondary School Students
Babiar, Tasha Calvert – Journal of Applied Measurement, 2011
Traditionally, women and minorities have not been fully represented in science and engineering. Numerous studies have attributed these differences to gaps in science achievement as measured by various standardized tests. Rather than describe mean group differences in science achievement across multiple cultures, this study focused on an in-depth…
Descriptors: Test Bias, Science Achievement, Standardized Tests, Grade 8

Ercikan, Kadriye – International Journal of Testing, 2002
Disentangled sources of differential item functioning (DIF) in a multilanguage assessment for which multiple factors were expected to be causing DIF. Data for the Third International Mathematics and Science study for four countries and two languages (3,000 to 11,000 cases in each comparison group) reveal amounts and sources of DIF. (SLD)
Descriptors: Cross Cultural Studies, English, French, International Studies
Sireci, Stephen G.; Gonzalez, Eugenio J. – 2003
International comparative educational studies make use of test instruments originally developed in English by international panels of experts, but that are ultimately administered in the language of instruction of the students. The comparability of the different language versions of these assessments is a critical issue in validating the…
Descriptors: Academic Achievement, Comparative Analysis, Difficulty Level, International Education
Learning about Students' Knowledge and Thinking in Science through Large-Scale Quantitative Studies.

Olsen, Rolf V.; Turmo, Are; Lie, Svein – European Journal of Psychology of Education, 2001
Discusses how responses to multiple-choice items could be interpreted, demonstrates how responses on constructed-response items can be analyzed, and examines interactions between item characteristics and student responses. Uses information, specifically items and student responses, from the Third International Mathematics and Science Study…
Descriptors: Educational Research, Higher Education, Mathematics Education, Science Education
Previous Page | Next Page »
Pages: 1 | 2