Publication Date
| In 2026 | 2 |
| Since 2025 | 462 |
| Since 2022 (last 5 years) | 1941 |
| Since 2017 (last 10 years) | 4513 |
| Since 2007 (last 20 years) | 6998 |
Descriptor
| Test Reliability | 15036 |
| Test Validity | 10004 |
| Test Construction | 4369 |
| Foreign Countries | 3831 |
| Psychometrics | 2428 |
| Factor Analysis | 2301 |
| Measures (Individuals) | 1785 |
| Evaluation Methods | 1410 |
| Higher Education | 1391 |
| Questionnaires | 1261 |
| Factor Structure | 1248 |
| More ▼ | |
Source
Author
Publication Type
Education Level
Audience
| Researchers | 454 |
| Practitioners | 319 |
| Teachers | 128 |
| Administrators | 73 |
| Policymakers | 33 |
| Counselors | 31 |
| Students | 17 |
| Parents | 10 |
| Community | 6 |
| Support Staff | 5 |
Location
| Turkey | 838 |
| Australia | 239 |
| China | 211 |
| Canada | 207 |
| Indonesia | 162 |
| Spain | 129 |
| United States | 123 |
| United Kingdom | 121 |
| Germany | 112 |
| Taiwan | 108 |
| Netherlands | 102 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 2 |
| Meets WWC Standards with or without Reservations | 2 |
| Does not meet standards | 1 |
Kleinke, David J. – 1976
Data from 200 college-level tests were used to compare three reliability approximations (two of Saupe and one of Cureton) to Kuder-Richardson Formula 20 (KR20). While the approximations correlated highly (about .9) with the reliability estimate, they tended to be underapproximations. The explanation lies in an apparent bias of Lord's approximation…
Descriptors: Comparative Analysis, Correlation, Error of Measurement, Statistical Analysis
Fitzgibbon, Thomas J. – 1975
In this speech given at the 1976 annual breakfast of the National Council on Measurement in Education Dr. Thomas J. Fitzgibbon, outgoing president of NCME, responds to critics of standardized testing and outlines the correct uses for it. He believes that many criticisms of standardized testing are due to misunderstanding or a lack of information.…
Descriptors: Achievement Tests, Standardized Tests, Test Reliability, Test Validity
Peer reviewedMays, Robert – Educational and Psychological Measurement, 1978
A FORTRAN program for clustering variables using the alpha coefficient of reliability is described. For batch operation, a rule for stopping the agglomerative precedure is available. The conversational version of the program allows the user to intervene in the process in order to test the final solution for sensitivity to changes. (Author/JKS)
Descriptors: Cluster Analysis, Computer Programs, Factor Analysis, Online Systems
Peer reviewedCicchetti, Domenic V.; And Others – Educational and Psychological Measurement, 1978
This program computes specific category agreement levels for both nominally and ordinally scaled data. For ordinally scaled data, an option is available for collapsing the original scale to a smaller number of categories, with the goal of improving the level of interrater reliability for the rating scale. (Author)
Descriptors: Attitude Measures, Computer Programs, Measurement Techniques, Rating Scales
Jones, Kenneth Melvin; Randolph, Daniel Lee – Southern Journal of Educational Research, 1978
Designed to measure self-actualization as defined by Abraham Maslow, the Jones Self Actualizing Scale, as assessed in this study, possesses content validity, reliability, and a number of other positive characteristics. (JC)
Descriptors: Measurement Instruments, Self Actualization, Test Construction, Test Reliability
Peer reviewedFleiss, Joseph L.; Shrout, Patrick E. – Psychometrika, 1978
When raters in a reliability study are a random sample, inferences about the intraclass correlation coefficient must be based on three mean squares from the analysis of variance: between subjects, between raters, and error. An approximate confidence interval for the parameter is presented as a function of these mean squares. (Author/JKS)
Descriptors: Analysis of Variance, Correlation, Measurement Techniques, Rating Scales
Peer reviewedGreene, Vernon L. – Psychometrika, 1978
The relationships among Rao's canonical factor analysis, Kaiser and Coffrey's alpha factor analysis, and Cronbach's alpha coefficient are discussed. Implications for test reliability and validity are noted. (JKS)
Descriptors: Correlation, Factor Analysis, Measurement Techniques, Psychometrics
Peer reviewedHassett, John J. – English Journal, 1978
Suggests four questions teachers can ask to determine whether a standardized test has provided accurate information about a particular student. (DD)
Descriptors: English Instruction, Secondary Education, Standardized Tests, Test Reliability
Peer reviewedGreen, Samual B.; And Others – Educational and Psychological Measurement, 1977
Confusion in the literature between the concepts of internal consistency and homogeneity has led to a misuse of coefficient alpha as an index of item homogeneity. This misuse is discussed and several indices of item homogeneity derived from the model of common factor analysis are offered as alternatives. (Author/JKS)
Descriptors: Factor Analysis, Item Analysis, Test Interpretation, Test Items
Fahey, Thomas D.; And Others – Research Quarterly, 1977
Results from comparative testing indicate that fingertip blood is a valid indicator of antecubital venous hematocrit (hct) and hemoglobin (hgb), and that hct ratios determined on the Coulter counter are comparable to those found by the microhematocrit method. (MB)
Descriptors: Blood Circulation, Chemical Analysis, Medical Research, Methods Research
Peer reviewedSamejima, Fumiko – Psychometrika, 1977
A method of estimating item characteristic functions is proposed, in which a set of test items, whose operating characteristics are known and which give a constant test information function for a wide range of ability, are used. The method is based on maximum likelihood estimation procedures. (Author/JKS)
Descriptors: Item Analysis, Latent Trait Theory, Measurement, Test Construction
Peer reviewedSubkoviak, Michael J.; Levin, Joel R. – Journal of Educational Measurement, 1977
Measurement error in dependent variables reduces the power of statistical tests to detect mean differences of specified magnitude. Procedures for determining power and sample size that consider the reliability of the dependent variable are discussed and illustrated. Methods for estimating reliability coefficients used in these procedures are…
Descriptors: Error of Measurement, Hypothesis Testing, Power (Statistics), Sampling
Peer reviewedWishart, Jennifer G. – American Journal of Mental Deficiency, 1987
Twelve children with Down's syndrome (ages 3-5 years) were tested six times over 2.5 months on three Piagetian infant search tasks. Results suggested that cognitive ability of this population may be poorly measured by single-session testing and that caution is necessary when using tests designed for and validated on younger, nonretarded subjects.…
Descriptors: Cognitive Measurement, Downs Syndrome, Test Reliability, Testing Problems
Peer reviewedMillsap, Roger E. – Applied Psychological Measurement, 1988
Two new methods for constructing a credibility interval (CI)--an interval containing a specified proportion of true validity description--are discussed, from a frequentist perspective. Tolerance intervals, unlike the current method of constructing the CI, have performance characteristics across repeated applications and may be useful in validity…
Descriptors: Bayesian Statistics, Meta Analysis, Statistical Analysis, Test Reliability
Chng, Chwee Lye; Roddy, William Meyer – Health Education (Washington D.C.), 1987
This article discusses the ethical implications of using tests intended only to protect the blood supply as diagnostic or prognostic tests for Acquired Immune Deficiency Syndrome (AIDS). The reliability of these tests and the confidentiality of their results are discussed, along with ethical issues of research regarding AIDS. (MT)
Descriptors: Acquired Immunodeficiency Syndrome, Ethics, Screening Tests, Social Discrimination


