NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20260
Since 20252
Since 2022 (last 5 years)4
Since 2017 (last 10 years)5
Since 2007 (last 20 years)10
Audience
Laws, Policies, & Programs
No Child Left Behind Act 20011
What Works Clearinghouse Rating
Showing all 10 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Owen Henkel; Hannah Horne-Robinson; Libby Hills; Bill Roberts; Josh McGrane – International Journal of Artificial Intelligence in Education, 2025
This paper reports on a set of three recent experiments utilizing large-scale speech models to assess the oral reading fluency (ORF) of students in Ghana. While ORF is a well-established measure of foundational literacy, assessing it typically requires one-on-one sessions between a student and a trained rater, a process that is time-consuming and…
Descriptors: Foreign Countries, Oral Reading, Reading Fluency, Literacy
Peer reviewed Peer reviewed
Direct linkDirect link
Paul A. Jewsbury; Matthew S. Johnson – Large-scale Assessments in Education, 2025
The standard methodology for many large-scale assessments in education involves regressing latent variables on numerous contextual variables to estimate proficiency distributions. To reduce the number of contextual variables used in the regression and improve estimation, we propose and evaluate principal component analysis on the covariance matrix…
Descriptors: Factor Analysis, Matrices, Regression (Statistics), Educational Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Suto, Irenka; Williamson, Joanna; Ireland, Jo; Macinska, Sylwia – Research Papers in Education, 2023
Errors that occasionally manifest in examination papers and other educational assessment instruments can threaten reliability and validity. For example, a multiple choice question could have two correct response options, or a geography question containing an inaccurate map could be unanswerable. In this paper we explore this oft-neglected element…
Descriptors: Error Patterns, International Assessment, Test Construction, Failure
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Karakaya, Ferhat; Yilmaz, Mehmet – Journal of Pedagogical Research, 2022
There have been significant advances in science and technology in recent years. Therefore, all countries need qualified people who can take on the challenges of life today and compete in the international arena. This has led countries to adopt new approaches to education. STEM education is one of the latest examples of those approaches. This study…
Descriptors: Science Teachers, Teacher Attitudes, Evaluation Methods, STEM Education
Peer reviewed Peer reviewed
Direct linkDirect link
Kalinowski, Steven T. – Educational and Psychological Measurement, 2019
Item response theory (IRT) is a statistical paradigm for developing educational tests and assessing students. IRT, however, currently lacks an established graphical method for examining model fit for the three-parameter logistic model, the most flexible and popular IRT model in educational testing. A method is presented here to do this. The graph,…
Descriptors: Item Response Theory, Educational Assessment, Goodness of Fit, Probability
Peer reviewed Peer reviewed
Direct linkDirect link
Noble, Tracy; Rosebery, Ann; Suarez, Catherine; Warren, Beth; O'Connor, Mary Catherine – Applied Measurement in Education, 2014
English language learners (ELLs) and their teachers, schools, and communities face increasingly high-stakes consequences due to test score gaps between ELLs and non-ELLs. It is essential that the field of educational assessment continue to investigate the meaning of these test score gaps. This article discusses the findings of an exploratory study…
Descriptors: English Language Learners, Evidence, Educational Assessment, Achievement Gap
Peer reviewed Peer reviewed
Direct linkDirect link
Gardner, John – Oxford Review of Education, 2013
Evidence from recent research suggests that in the UK the public perception of errors in national examinations is that they are simply mistakes; events that are preventable. This perception predominates over the more sophisticated technical view that errors arise from many sources and create an inevitable variability in assessment outcomes. The…
Descriptors: Educational Assessment, Public Opinion, Error of Measurement, Foreign Countries
Peer reviewed Peer reviewed
Direct linkDirect link
Shortridge, Ashton; Goldsberry, Kirk; Weessies, Kathleen – Journal of Geography, 2011
This article characterizes and measures errors in the 2010 National Research Council (NRC) assessment of research-doctorate programs in geography. This article provides a conceptual model for data-based sources of uncertainty and reports on a quantitative assessment of NRC research data uncertainty for a particular geography doctoral program.…
Descriptors: Geography, Doctoral Programs, Graduate Study, Educational Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Wyse, Adam E.; Mapuranga, Raymond – International Journal of Testing, 2009
Differential item functioning (DIF) analysis is a statistical technique used for ensuring the equity and fairness of educational assessments. This study formulates a new DIF analysis method using the information similarity index (ISI). ISI compares item information functions when data fits the Rasch model. Through simulations and an international…
Descriptors: Test Bias, Evaluation Methods, Test Items, Educational Assessment
Harris, Douglas N. – Policy Analysis for California Education, PACE (NJ3), 2010
In this policy brief, the author explores the problems with attainment measures when it comes to evaluating performance at the school level, and explores the best uses of value-added measures. These value-added measures, the author writes, are useful for sorting out-of-school influences from school influences or from teacher performance, giving…
Descriptors: Principals, Observation, Teacher Evaluation, Measurement Techniques