NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1 to 15 of 76 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Hyo Jeong Shin; Christoph König; Frederic Robin; Andreas Frey; Kentaro Yamamoto – Journal of Educational Measurement, 2025
Many international large-scale assessments (ILSAs) have switched to multistage adaptive testing (MST) designs to improve measurement efficiency in measuring the skills of the heterogeneous populations around the world. In this context, previous literature has reported the acceptable level of model parameter recovery under the MST designs when the…
Descriptors: Robustness (Statistics), Item Response Theory, Adaptive Testing, Test Construction
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Reckase, Mark D. – ETS Research Report Series, 2017
A common interpretation of achievement test results is that they provide measures of achievement that are much like other measures we commonly use for height, weight, or the cost of goods. In a limited sense, such interpretations are correct, but some nuances of these interpretations have important implications for the use of achievement test…
Descriptors: Models, Achievement Tests, Test Results, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Robitzsch, Alexander; Lüdtke, Oliver – Large-scale Assessments in Education, 2023
One major aim of international large-scale assessments (ILSA) like PISA is to monitor changes in student performance over time. To accomplish this task, a set of common items (i.e., link items) is repeatedly administered in each assessment. Linking methods based on item response theory (IRT) models are used to align the results from the different…
Descriptors: Educational Trends, Trend Analysis, International Assessment, Achievement Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Hickey, Daniel; Harris, Tripp – Distance Education, 2021
Increased online learning is helping many appreciate that online grading, formative assessment, and summative testing can cause instructor burnout and leave little time for more productive instructor interactions. We reimagined grading, assessment, and testing in an extended program of design-based research using situative theory to refine online…
Descriptors: Computer Assisted Testing, Grading, Student Evaluation, Online Courses
Peer reviewed Peer reviewed
Direct linkDirect link
Ma, Wenchao; de la Torre, Jimmy – Journal of Educational and Behavioral Statistics, 2019
Solving a constructed-response item usually requires successfully performing a sequence of tasks. Each task could involve different attributes, and those required attributes may be "condensed" in various ways to produce the responses. The sequential generalized deterministic input noisy "and" gate model is a general cognitive…
Descriptors: Test Items, Cognitive Measurement, Models, Hypothesis Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Wise, Steven L.; Kingsbury, G. Gage – Journal of Educational Measurement, 2016
This study examined the utility of response time-based analyses in understanding the behavior of unmotivated test takers. For the data from an adaptive achievement test, patterns of observed rapid-guessing behavior and item response accuracy were compared to the behavior expected under several types of models that have been proposed to represent…
Descriptors: Achievement Tests, Student Motivation, Test Wiseness, Adaptive Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Forestier, Katherine; Adamson, Bob – Compare: A Journal of Comparative and International Education, 2017
Adopting a comparative perspective to address educational issues in different contexts was a hallmark of Jullien's work in the early-nineteenth century. Different emphases and approaches to comparative education methodology have emerged in recent times thanks to major developments in technology, but have these changes rendered Jullien's ideas…
Descriptors: Criticism, International Assessment, Comparative Education, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Debeer, Dries; Janssen, Rianne; De Boeck, Paul – Journal of Educational Measurement, 2017
When dealing with missing responses, two types of omissions can be discerned: items can be skipped or not reached by the test taker. When the occurrence of these omissions is related to the proficiency process the missingness is nonignorable. The purpose of this article is to present a tree-based IRT framework for modeling responses and omissions…
Descriptors: Item Response Theory, Test Items, Responses, Testing Problems
Peer reviewed Peer reviewed
Direct linkDirect link
Sideridis, Georgios; Tsaousis, Ioannis; Al Harbi, Khaleel – Educational and Psychological Measurement, 2017
The purpose of the present article was to illustrate, using an example from a national assessment, the value from analyzing the behavior of distractors in measures that engage the multiple-choice format. A secondary purpose of the present article was to illustrate four remedial actions that can potentially improve the measurement of the…
Descriptors: Multiple Choice Tests, Attention Control, Testing, Remedial Instruction
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Nadrah; Tolla, Ismail; Ali, Muhammad Sidin; Muris – International Education Studies, 2017
This research aims at describing the effect of cooperative learning model of Teams Games Tournament (TGT) and motivation toward physics learning outcome. This research was a quasi-experimental research with a factorial design conducted at SMAN 2 Makassar. Independent variables were learning models. They were cooperative learning model of TGT and…
Descriptors: Foreign Countries, Cooperative Learning, Models, Educational Games
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Chen, Huilin – Journal of Education and Learning, 2014
The validity of the computer-based language test is possibly affected by three factors: computer familiarity, audio-visual cognitive competence, and other discrepancies in construct. Therefore, validating the equivalence between the paper-and-pencil language test and the computer-based language test is a key step in the procedure of designing a…
Descriptors: Computer Assisted Testing, Language Tests, Test Validity, Case Studies
Peer reviewed Peer reviewed
Direct linkDirect link
Li, Ying; Jiao, Hong; Lissitz, Robert W. – Journal of Applied Testing Technology, 2012
This study investigated the application of multidimensional item response theory (IRT) models to validate test structure and dimensionality. Multiple content areas or domains within a single subject often exist in large-scale achievement tests. Such areas or domains may cause multidimensionality or local item dependence, which both violate the…
Descriptors: Achievement Tests, Science Tests, Item Response Theory, Measures (Individuals)
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Hansen, Mark; Cai, Li; Monroe, Scott; Li, Zhen – Grantee Submission, 2016
Despite the growing popularity of diagnostic classification models (e.g., Rupp, Templin, & Henson, 2010) in educational and psychological measurement, methods for testing their absolute goodness-of-fit to real data remain relatively underdeveloped. For tests of reasonable length and for realistic sample size, full-information test statistics…
Descriptors: Goodness of Fit, Item Response Theory, Classification, Maximum Likelihood Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Krskova, Hana; Baumann, Chris – International Journal of Educational Management, 2017
Purpose: The purpose of this paper is to combine seemingly unrelated factors to explain global competitiveness. The study argues that school discipline and education investment affect competitiveness with the association being mediated by educational performance. Crucially, diachronic effects of discipline on performance are tested to demonstrate…
Descriptors: Foreign Countries, Competition, Academic Achievement, Least Squares Statistics
Brown, James Dean; Salmani Nodoushan, Mohammad Ali – Online Submission, 2015
In this interview, JD Brown reflects on language testing/assessment. He suggests that language testing can be seen as a continuum with hard core positivist approaches at one end and post modernist interpretive perspectives at the other, and also argues that norm referencing (be it proficiency, placement, or aptitude testing) and criterion…
Descriptors: Interviews, Language Tests, English (Second Language), Second Language Learning
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5  |  6