Publication Date
| In 2026 | 0 |
| Since 2025 | 53 |
| Since 2022 (last 5 years) | 195 |
| Since 2017 (last 10 years) | 495 |
| Since 2007 (last 20 years) | 743 |
Descriptor
| Test Items | 1187 |
| Test Reliability | 1187 |
| Test Validity | 685 |
| Test Construction | 566 |
| Foreign Countries | 349 |
| Difficulty Level | 280 |
| Item Analysis | 253 |
| Psychometrics | 234 |
| Item Response Theory | 219 |
| Factor Analysis | 183 |
| Multiple Choice Tests | 173 |
| More ▼ | |
Source
Author
| Schoen, Robert C. | 12 |
| LaVenia, Mark | 5 |
| Liu, Ou Lydia | 5 |
| Anderson, Daniel | 4 |
| Bauduin, Charity | 4 |
| DiLuzio, Geneva J. | 4 |
| Farina, Kristy | 4 |
| Haladyna, Thomas M. | 4 |
| Huck, Schuyler W. | 4 |
| Petscher, Yaacov | 4 |
| Stansfield, Charles W. | 4 |
| More ▼ | |
Publication Type
Education Level
Audience
| Practitioners | 39 |
| Researchers | 30 |
| Teachers | 24 |
| Administrators | 13 |
| Support Staff | 3 |
| Counselors | 2 |
| Students | 2 |
| Community | 1 |
| Parents | 1 |
| Policymakers | 1 |
Location
| Turkey | 69 |
| Indonesia | 37 |
| Germany | 20 |
| Canada | 17 |
| Florida | 17 |
| China | 16 |
| Australia | 15 |
| California | 12 |
| Iran | 11 |
| India | 10 |
| New York | 9 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 1 |
| Meets WWC Standards with or without Reservations | 1 |
Ashvind Nand Singh – ProQuest LLC, 2008
Due to the relative inability of individuals with intellectual disabilities (ID) to provide an accurate and reliable self-report, assessment in this population is more difficult than with individuals in the general population. As such, assessment procedures must be adjusted to compensate for the relative lack of information that the individual can…
Descriptors: Test Items, Item Analysis, Test Construction, Behavior Rating Scales
Marshall, Robert C.; Wright, Heather Harris – American Journal of Speech-Language Pathology, 2007
Purpose: The Kentucky Aphasia Test (KAT) is an objective measure of language functioning for persons with aphasia. This article describes materials, administration, and scoring of the KAT; presents the rationale for development of test items; reports information from a pilot study; and discusses the role of the KAT in aphasia assessment. Method:…
Descriptors: Aphasia, Test Format, Language Tests, Expressive Language
Braden, Jeffery P.; Iribarren, Jacqueline A. – Journal of Psychoeducational Assessment, 2007
In this article, the authors review the Wechsler Intelligence Scale for Children-Fourth Edition Spanish (WISC-IV Spanish), a Spanish translation and adaptation of the WISC-IV. The test was developed to measure the intellectual ability of Spanish-speaking children in the United States ages 6 years, 0 months, through 16 years, 11 months. These…
Descriptors: Intelligence Tests, Spanish, Translation, Children
Bernard, Larry C.; Mills, Michael; Swenson, Leland; Walsh, R. Patricia – Assessment, 2008
We report the development of the Assessment of Individual Motives-Questionnaire (AIM-Q), a new instrument based on an evolutionary psychology theory of human motivation. It provides multitrait-multimethod (MTMM) assessment of individual differences on 15 motive scales. A total heterogeneous sample of N = 1,251 participated in eight studies that…
Descriptors: Test Construction, Questionnaires, Test Reliability, Multitrait Multimethod Techniques
Chavez, Oscar; Papick, Ira; Ross, Dan J.; Grouws, Douglas A. – Online Submission, 2010
The purpose of this paper was to describe the process of development of assessment instruments for the Comparing Options in Secondary Mathematics: Investigating Curriculum (COSMIC) project. The COSMIC project was a three-year longitudinal comparative study focusing on evaluating high school students' mathematics learning from two distinct…
Descriptors: Mathematics Education, Mathematics Achievement, Interrater Reliability, Scoring Rubrics
Peer reviewedCliff, Norman – Journal of Educational Statistics, 1984
The proposed coefficient is derived by assuming that the average Goodman-Kruskal gamma between items of identical difficulty would be the same for items of different difficulty. An estimate of covariance between items of identical difficulty leads to an estimate of the correlation between two tests with identical distributions of difficulty.…
Descriptors: Difficulty Level, Mathematical Formulas, Test Items, Test Reliability
Hall, John D.; Howerton, D. Lynn; Jones, Craig H. – Research in the Schools, 2008
The No Child Left Behind Act and the accountability movement in public education caused many states to develop criterion-referenced academic achievement tests. Scores from these tests are often used to make high stakes decisions. Even so, these tests typically do not receive independent psychometric scrutiny. We evaluated the 2005 Arkansas…
Descriptors: Criterion Referenced Tests, Achievement Tests, High Stakes Tests, Public Education
Chiat, Shula; Roy, Penny – Journal of Speech, Language, and Hearing Research, 2007
Purpose: To determine the psychometric properties of the Preschool Repetition (PSRep) Test (Roy & Chiat, 2004), to establish the range of performance in typically developing children and variables affecting this performance, and to compare the performance of clinically referred children. Method: The PSRep Test comprises 18 words and 18…
Descriptors: Phonology, Psychometrics, Interrater Reliability, Followup Studies
Lee, Sang Min; Puig, Ana; Pasquarella-Daley, Lauren; Denny, George; Rai, Ann Allen; Dallape, Aprille; Parker, Woodrow Max – Measurement and Evaluation in Counseling and Development, 2007
This article describes the revision of the White Racial Consciousness Development Scale (D. Claney & W. M. Parker, 1989). A multistage approach including item generation, item refinement and selection, and evaluation of score validity and reliability was used to test construction and validation. Implications for theory, practice, and future…
Descriptors: Measures (Individuals), Test Construction, Test Items, Scores
Pepin, Michel – 1983
This paper presents three different ways of computing the internal consistency coefficient alpha for a same set of data. The main objective of the paper is the illustration of a method for maximizing coefficient alpha. The maximization of alpha can be achieved with the aid of a principal component analysis. The relation between alpha max. and the…
Descriptors: Research Methodology, Research Problems, Statistical Analysis, Test Items
Peer reviewedGreen, Samual B.; And Others – Educational and Psychological Measurement, 1977
Confusion in the literature between the concepts of internal consistency and homogeneity has led to a misuse of coefficient alpha as an index of item homogeneity. This misuse is discussed and several indices of item homogeneity derived from the model of common factor analysis are offered as alternatives. (Author/JKS)
Descriptors: Factor Analysis, Item Analysis, Test Interpretation, Test Items
Peer reviewedShapiro, Alexander – Psychometrika, 1982
Minimum trace factor analysis has been used to find the greatest lower bound to reliability. This technique, however, fails to be scale free. A solution to the scale problem is proposed through the maximization of the greatest lower bound as the function of weights. (Author/JKS)
Descriptors: Algorithms, Estimation (Mathematics), Factor Analysis, Psychometrics
Peer reviewedRaju, Nambury S. – Psychometrika, 1979
An important relationship is given for two generalizations of coefficient alpha: (1) Rajaratnam, Cronbach, and Gleser's generalizability formula for stratified-parallel tests, and (2) Raju's coefficient beta. (Author/CTM)
Descriptors: Item Analysis, Mathematical Formulas, Test Construction, Test Items
Peer reviewedCudeck, Robert – Journal of Educational Measurement, 1980
Methods for evaluating the consistency of responses to test items were compared. When a researcher is unwilling to make the assumptions of classical test theory, has only a small number of items, or is in a tailored testing context, Cliff's dominance indices may be useful. (Author/CTM)
Descriptors: Error Patterns, Item Analysis, Test Items, Test Reliability
Peer reviewedFox, Robert A. – Journal of School Health, 1980
Some practical guidelines for developing multiple choice tests are offered. Included are three steps: (1) test design; (2) proper construction of test items; and (3) item analysis and evaluation. (JMF)
Descriptors: Guidelines, Objective Tests, Planning, Test Construction

Direct link
