Publication Date
In 2025 | 37 |
Since 2024 | 146 |
Since 2021 (last 5 years) | 509 |
Since 2016 (last 10 years) | 1088 |
Since 2006 (last 20 years) | 2416 |
Descriptor
Item Analysis | 3104 |
Foreign Countries | 1096 |
Test Items | 926 |
Factor Analysis | 681 |
Psychometrics | 604 |
Measures (Individuals) | 507 |
Correlation | 504 |
Test Validity | 484 |
Test Construction | 479 |
Test Reliability | 426 |
Item Response Theory | 421 |
More ▼ |
Source
Author
Publication Type
Education Level
Location
Turkey | 169 |
China | 68 |
Australia | 59 |
Canada | 53 |
Germany | 42 |
Taiwan | 40 |
United States | 38 |
United Kingdom | 37 |
Iran | 36 |
Japan | 33 |
Spain | 31 |
More ▼ |
Laws, Policies, & Programs
No Child Left Behind Act 2001 | 14 |
Individuals with Disabilities… | 5 |
Deferred Action for Childhood… | 1 |
Every Student Succeeds Act… | 1 |
Individuals with Disabilities… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Does not meet standards | 1 |
Pere J. Ferrando; David Navarro-González; Urbano Lorenzo-Seva – Educational and Psychological Measurement, 2024
Descriptive fit indices that do not require a formal statistical basis and do not specifically depend on a given estimation criterion are useful as auxiliary devices for judging the appropriateness of unrestricted or exploratory factor analytical (UFA) solutions, when the problem is to decide the most appropriate number of common factors. While…
Descriptors: Factor Analysis, Item Analysis, Effect Size, Goodness of Fit
Miguel A. García-Pérez – Educational and Psychological Measurement, 2024
A recurring question regarding Likert items is whether the discrete steps that this response format allows represent constant increments along the underlying continuum. This question appears unsolvable because Likert responses carry no direct information to this effect. Yet, any item administered in Likert format can identically be administered…
Descriptors: Likert Scales, Test Construction, Test Items, Item Analysis
Sohee Kim; Ki Lynn Cole – International Journal of Testing, 2025
This study conducted a comprehensive comparison of Item Response Theory (IRT) linking methods applied to a bifactor model, examining their performance on both multiple choice (MC) and mixed format tests within the common item nonequivalent group design framework. Four distinct multidimensional IRT linking approaches were explored, consisting of…
Descriptors: Item Response Theory, Comparative Analysis, Models, Item Analysis
Chan Zhang; Shuaiying Cao; Minglei Wang; Jiangyan Wang; Lirui He – Field Methods, 2025
Previous research on grid questions has mostly focused on their comparability with the item-by-item method and the use of shading to help respondents navigate through a grid. This study extends prior work by examining whether lexical similarity among grid items affects how respondents answer the questions in an experiment where we manipulated…
Descriptors: Foreign Countries, Surveys, Test Construction, Design
Gregory H. Peterson; Michael B. Kozlowski – Measurement and Evaluation in Counseling and Development, 2024
This study aimed to develop a scale to assess counselors' ability to provide counseling to address the mental health impacts of climate change. Over three studies, we provide reliability and validity evidence for a Climate Change Counseling Scale (3CS) in a large representative sample of counselors across the US. In study one and two, an…
Descriptors: Counselors, Mental Health, Climate, Test Construction
Bronson Hui; Zhiyi Wu – Studies in Second Language Acquisition, 2024
A slowdown or a speedup in response times across experimental conditions can be taken as evidence of online deployment of knowledge. However, response-time difference measures are rarely evaluated on their reliability, and there is no standard practice to estimate it. In this article, we used three open data sets to explore an approach to…
Descriptors: Reliability, Reaction Time, Psychometrics, Criticism
Cindy M. Gilson; Claudia Flowers; Wen-Hsuan Chang – International Journal of Listening, 2024
Effective teacher listening is critical in establishing a culturally responsive and equitable learning environment for all students. Teacher listening orientations are ways in which teachers are situated to listen to students. Measurements of their orientations during classroom discourse is an important yet under-researched tool for enhancing both…
Descriptors: Listening, Dialogs (Language), Teachers, Measurement Techniques
Xiaowen Liu – International Journal of Testing, 2024
Differential item functioning (DIF) often arises from multiple sources. Within the context of multidimensional item response theory, this study examined DIF items with varying secondary dimensions using the three DIF methods: SIBTEST, Mantel-Haenszel, and logistic regression. The effect of the number of secondary dimensions on DIF detection rates…
Descriptors: Item Analysis, Test Items, Item Response Theory, Correlation
Gal Kaldes; Elizabeth L. Tighe – Reading and Writing: An Interdisciplinary Journal, 2024
The purpose of this study was to (a) examine the underlying assessment structure of the Derivational Morphology Task (DMORPH) and (b) investigate the relation of the DMORPH to vocabulary and reading comprehension outcomes with a sample of struggling adult readers. Specifically, participants included 218 struggling adult readers enrolled in adult…
Descriptors: Adult Literacy, Reading Tests, Reading Comprehension, Phonological Awareness
Karl Schweizer; Andreas Gold; Dorothea Krampen; Stefan Troche – Educational and Psychological Measurement, 2024
Conceptualizing two-variable disturbances preventing good model fit in confirmatory factor analysis as item-level method effects instead of correlated residuals avoids violating the principle that residual variation is unique for each item. The possibility of representing such a disturbance by a method factor of a bifactor measurement model was…
Descriptors: Correlation, Factor Analysis, Measurement Techniques, Item Analysis
Laurie Graham Dodge – Assessment Update, 2024
The importance of assessment has increased and is now integrated in institutional cultures for faculty, leaders, service units, accreditors, and learners. To address the call for validation of learning, institutions and organizations are building competency-based education (CBE) programs that are relevant, accessible, and ensure quality. In 2021,…
Descriptors: Evaluation, Job Skills, Competency Based Education, Learner Controlled Instruction
Martin Bäckström; Fredrik Björklund – Educational and Psychological Measurement, 2024
The forced-choice response format is often considered superior to the standard Likert-type format for controlling social desirability in personality inventories. We performed simulations and found that the trait information based on the two formats converges when the number of items is high and forced-choice items are mixed with regard to…
Descriptors: Likert Scales, Item Analysis, Personality Traits, Personality Measures
Olajumoke Beulah Adigun – Learning Environments Research, 2024
Particularly for adolescent students, tools for measuring their experience of negatively consequential aspects of student-teacher relationships are limited even though evidence shows that keeping this group motivated in the classroom is especially challenging. Therefore, using basic psychological needs theory, this paper proposes a…
Descriptors: Adolescents, Teacher Student Relationship, Educational Environment, Psychological Needs
Moore, C. Missy; Mullen, Patrick R.; Hinchey, Kaitlin J.; Lambie, Glenn W. – Counselor Education and Supervision, 2023
Our study examines the differential item functioning of the Counselor Competencies Scale--Revised (CCS-R) scores due to respondents' gender, the type of evaluation, and a combination of these two variables using a large sample (N = 1614). Implications of the findings are offered to inform counselor educators and supervisors using the CCS-R and…
Descriptors: Item Analysis, Measures (Individuals), Counselors, Competence
Kazuhiro Yamaguchi – Journal of Educational and Behavioral Statistics, 2025
This study proposes a Bayesian method for diagnostic classification models (DCMs) for a partially known Q-matrix setting between exploratory and confirmatory DCMs. This Q-matrix setting is practical and useful because test experts have pre-knowledge of the Q-matrix but cannot readily specify it completely. The proposed method employs priors for…
Descriptors: Models, Classification, Bayesian Statistics, Evaluation Methods