Publication Date
| In 2026 | 0 |
| Since 2025 | 1 |
| Since 2022 (last 5 years) | 4 |
| Since 2017 (last 10 years) | 7 |
| Since 2007 (last 20 years) | 18 |
Descriptor
| Error Patterns | 27 |
| Evaluation Methods | 27 |
| Computation | 6 |
| Comparative Analysis | 5 |
| Educational Assessment | 4 |
| Feedback (Response) | 4 |
| Item Response Theory | 4 |
| Models | 4 |
| Student Evaluation | 4 |
| Test Items | 4 |
| Data Analysis | 3 |
| More ▼ | |
Source
Author
| An, Min | 1 |
| Attali, Yigal | 1 |
| Bird, Kevin D. | 1 |
| Broen, Patricia A. | 1 |
| Brown, Scott D. | 1 |
| Carlos Alario-Hoyos | 1 |
| Carlos Delgado Kloos | 1 |
| Casey, Sean D. | 1 |
| Chan, Daniel W.-L. | 1 |
| Chan, Wai | 1 |
| Chitpin, Stephanie | 1 |
| More ▼ | |
Publication Type
| Reports - Evaluative | 27 |
| Journal Articles | 24 |
| Information Analyses | 1 |
| Opinion Papers | 1 |
| Reports - Descriptive | 1 |
| Speeches/Meeting Papers | 1 |
| Tests/Questionnaires | 1 |
Education Level
| Higher Education | 3 |
| Postsecondary Education | 2 |
| Elementary Secondary Education | 1 |
| Grade 4 | 1 |
| Grade 6 | 1 |
| Grade 8 | 1 |
| Secondary Education | 1 |
Audience
| Policymakers | 1 |
| Practitioners | 1 |
| Researchers | 1 |
| Teachers | 1 |
Location
| California | 1 |
| Georgia | 1 |
Laws, Policies, & Programs
| National Direct Student Loan… | 1 |
| No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
| Kaufman Test of Educational… | 1 |
| Program for International… | 1 |
What Works Clearinghouse Rating
Joo, Seang-Hwane; Lee, Philseok – Journal of Educational Measurement, 2022
Abstract This study proposes a new Bayesian differential item functioning (DIF) detection method using posterior predictive model checking (PPMC). Item fit measures including infit, outfit, observed score distribution (OSD), and Q1 were considered as discrepancy statistics for the PPMC DIF methods. The performance of the PPMC DIF method was…
Descriptors: Test Items, Bayesian Statistics, Monte Carlo Methods, Prediction
Suto, Irenka; Williamson, Joanna; Ireland, Jo; Macinska, Sylwia – Research Papers in Education, 2023
Errors that occasionally manifest in examination papers and other educational assessment instruments can threaten reliability and validity. For example, a multiple choice question could have two correct response options, or a geography question containing an inaccurate map could be unanswerable. In this paper we explore this oft-neglected element…
Descriptors: Error Patterns, International Assessment, Test Construction, Failure
Iria Estévez-Ayres; Patricia Callejo; Miguel Ángel Hombrados-Herrera; Carlos Alario-Hoyos; Carlos Delgado Kloos – International Journal of Artificial Intelligence in Education, 2025
The emergence of Large Language Models (LLMs) has marked a significant change in education. The appearance of these LLMs and their associated chatbots has yielded several advantages for both students and educators, including their use as teaching assistants for content creation or summarisation. This paper aims to evaluate the capacity of LLMs…
Descriptors: Artificial Intelligence, Natural Language Processing, Computer Mediated Communication, Technology Uses in Education
James Van Loon; Shawn McCann – Journal of Electronic Resources Librarianship, 2024
Academic libraries must provide accessible electronic and information technology to patrons using assistive technology to access library materials. Accessibility of scholarly material is particularly important in the STEM disciplines, where attrition rates of students with disabilities is higher than in other fields. In this study, the…
Descriptors: Academic Libraries, STEM Education, Accessibility (for Disabled), Bibliographic Databases
McCloskey, George – Journal of Psychoeducational Assessment, 2017
This commentary will take an historical perspective on the Kaufman Test of Educational Achievement (KTEA) error analysis, discussing where it started, where it is today, and where it may be headed in the future. In addition, the commentary will compare and contrast the KTEA error analysis procedures that are rooted in psychometric methodology and…
Descriptors: Achievement Tests, Error Patterns, Comparative Analysis, Psychometrics
Reed, Jolene B.; Kellum, Marcia L. – Texas Association for Literacy Education Yearbook, 2017
Running records are not regularly utilized to their full potential by classroom teachers. Often, teachers use running records only to determine an instructional text level for an individual student and then to place students into reading groups for instructional purposes based on student reading levels. When more thoroughly analyzed, running…
Descriptors: Literacy Education, Data Use, Decision Making, Teaching Methods
Yu, Chong Ho; Douglas, Samantha; Lee, Anna; An, Min – Practical Assessment, Research & Evaluation, 2016
This paper aims to illustrate how data visualization could be utilized to identify errors prior to modeling, using an example with multi-dimensional item response theory (MIRT). MIRT combines item response theory and factor analysis to identify a psychometric model that investigates two or more latent traits. While it may seem convenient to…
Descriptors: Visualization, Item Response Theory, Sample Size, Correlation
Chitpin, Stephanie – International Journal of Educational Management, 2017
Purpose: The purpose of this paper is to illustrate how associationism mistakenly assumes that direct experience is possible; that is, there is expectation-free observation and association without prior expectation. Thus, associationism assumes that learning involves the absorption of information from the environment itself. However, contrary…
Descriptors: Abstract Reasoning, Associative Learning, Association (Psychology), Philosophy
Hung, Lai-Fa – Applied Psychological Measurement, 2012
Rasch used a Poisson model to analyze errors and speed in reading tests. An important property of the Poisson distribution is that the mean and variance are equal. However, in social science research, it is very common for the variance to be greater than the mean (i.e., the data are overdispersed). This study embeds the Rasch model within an…
Descriptors: Social Science Research, Markov Processes, Reading Tests, Social Sciences
Bird, Kevin D. – Psychological Methods, 2011
Any set of confidence interval inferences on J - 1 linearly independent contrasts on J means, such as the two comparisons [mu][subscript 1] - [mu][subscript 2] and [mu][subscript 2] - [mu][subscript 3] on 3 means, provides a basis for the deduction of interval inferences on all other contrasts, such as the redundant comparison [mu][subscript 1] -…
Descriptors: Intervals, Statistical Analysis, Inferences, Comparative Analysis
Iamarino, Danielle L. – Current Issues in Education, 2014
This paper explores the methodology and application of an assessment philosophy known as standards-based grading, via a critical comparison of standards-based grading to other assessment philosophies commonly employed at the elementary, secondary, and post-secondary levels of education. Evidenced by examples of increased student engagement and…
Descriptors: Grading, Evaluation Methods, Evaluation Criteria, Evaluation Research
Attali, Yigal – Applied Psychological Measurement, 2011
Recently, Attali and Powers investigated the usefulness of providing immediate feedback on the correctness of answers to constructed response questions and the opportunity to revise incorrect answers. This article introduces an item response theory (IRT) model for scoring revised responses to questions when several attempts are allowed. The model…
Descriptors: Feedback (Response), Item Response Theory, Models, Error Correction
Puhan, Gautam – Applied Measurement in Education, 2009
The purpose of this study is to determine the extent of scale drift on a test that employs cut scores. It was essential to examine scale drift for this testing program because new forms in this testing program are often put on scale through a series of intermediate equatings (known as equating chains). This process may cause equating error to…
Descriptors: Testing Programs, Testing, Measurement Techniques, Item Response Theory
Wyse, Adam E.; Mapuranga, Raymond – International Journal of Testing, 2009
Differential item functioning (DIF) analysis is a statistical technique used for ensuring the equity and fairness of educational assessments. This study formulates a new DIF analysis method using the information similarity index (ISI). ISI compares item information functions when data fits the Rasch model. Through simulations and an international…
Descriptors: Test Bias, Evaluation Methods, Test Items, Educational Assessment
Harris, Douglas N. – Policy Analysis for California Education, PACE (NJ3), 2010
In this policy brief, the author explores the problems with attainment measures when it comes to evaluating performance at the school level, and explores the best uses of value-added measures. These value-added measures, the author writes, are useful for sorting out-of-school influences from school influences or from teacher performance, giving…
Descriptors: Principals, Observation, Teacher Evaluation, Measurement Techniques
Previous Page | Next Page »
Pages: 1 | 2
Peer reviewed
Direct link
