Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 0 |
| Since 2007 (last 20 years) | 6 |
Descriptor
Source
| Applied Psychological… | 72 |
Author
| Divgi, D. R. | 2 |
| Humphreys, Lloyd G. | 2 |
| Wang, Wen-Chung | 2 |
| Whitely, Susan E. | 2 |
| van der Linden, Wim J. | 2 |
| Backteman, G. | 1 |
| Baker, A. Harvey | 1 |
| Barnes, Janet L. | 1 |
| Bechger, Timo M. | 1 |
| Bejar, Isaac I. | 1 |
| Biswas, Ajoy Kumar | 1 |
| More ▼ | |
Publication Type
Education Level
| Higher Education | 1 |
| Postsecondary Education | 1 |
Audience
Location
| West Germany | 2 |
| Australia | 1 |
| Netherlands | 1 |
| Sweden | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Peer reviewedLuecht, Richard M. – Applied Psychological Measurement, 1996
The example of a medical licensure test is used to demonstrate situations in which complex, integrated content must be balanced at the total test level for validity reasons, but items assigned to reportable subscore categories may be used under a multidimensional item response theory adaptive paradigm to improve subscore reliability. (SLD)
Descriptors: Adaptive Testing, Certification, Computer Assisted Testing, Licensing Examinations (Professions)
Peer reviewedDavison, Mark L.; Robbins, Stephen – Applied Psychological Measurement, 1978
Empirically weighted scores for Rest's Defining Issues Test were found to be more reliable than the simple sum of scores theoretically weighted sum, or Rest's p scores. They also had slightly higher correlations with Kohlberg's interview scores. Empirically weighted scores also showed more significant change in two longitudinal studies. (CTM)
Descriptors: Higher Education, Longitudinal Studies, Moral Development, Moral Values
Peer reviewedPoizner, Sharon B.; And Others – Applied Psychological Measurement, 1978
Binary, probability, and ordinal scoring procedures for multiple-choice items were examined. In two situations, it was found that both the probability and ordinal scoring systems were more reliable than the binary scoring method. (Author/CTM)
Descriptors: Confidence Testing, Guessing (Tests), Higher Education, Multiple Choice Tests
Peer reviewedLunneborg, Clifford E. – Applied Psychological Measurement, 1977
Three studies are described in which choice reaction time (RT) was related to such psychometric ability measures as verbal comprehension, numerical reasoning, hidden figures, and progressive matrices tests. Fairly consistent negative correlations were found between these tests and choice RT when high school samples were used. (Author/CTM)
Descriptors: Cognitive Ability, Cognitive Processes, High Schools, Higher Education
Peer reviewedHambleton, Ronald K., Ed. – Applied Psychological Measurement, 1980
This special issue covers recent technical developments in the field of criterion-referenced testing. An introduction, six papers, and two commentaries dealing with test development, test score uses, and evaluation of scores review relevant literature, offer new models and/or results, and suggest directions for additional research. (SLD)
Descriptors: Criterion Referenced Tests, Mastery Tests, Measurement Techniques, Standard Setting (Scoring)
Peer reviewedFrederiksen, Norman; Ward, William C. – Applied Psychological Measurement, 1978
A set of Tests of Scientific Thinking were developed for possible use as criterion measures in research on creativity. Scores on the tests describe both quality and quantity of ideas produced in formulating hypotheses, evaluating proposals, solving methodological problems, and devising methods for measuring constructs. (Author/CTM)
Descriptors: Creativity Tests, Higher Education, Item Sampling, Predictive Validity
Peer reviewedSchmeck, Ronald Ray; And Others – Applied Psychological Measurement, 1977
Five studies are presented describing the development of a self-report inventory for measuring individual differences in learning processes. Factor analysis of items yielded four scales: Synthesis-Analysis, Study Methods, Fact Retention, and Elaborative Processing. There were no sex differences, and the scales demonstrated acceptable reliabilities…
Descriptors: Factor Analysis, Higher Education, Learning Processes, Retention (Psychology)
Peer reviewedHsu, Louis M. – Applied Psychological Measurement, 1979
A comparison of the relative ordering power of separate and grouped-items true-false tests indicated that neither type of test was uniformly superior to the other across all levels of knowledge of examinees. Grouped-item tests were found superior for examinees with low levels of knowledge. (Author/CTM)
Descriptors: Academic Ability, Knowledge Level, Multiple Choice Tests, Scores
Peer reviewedMagnusson, D.; Backteman, G. – Applied Psychological Measurement, 1979
A longitudinal study of approximately 1,000 students aged 10-16 showed high stability of intelligence and creativity. Stability coefficients for intelligence were higher than those for creativity. Results supported the construct validity of creativity. (MH)
Descriptors: Creativity, Creativity Tests, Elementary Secondary Education, Foreign Countries
Peer reviewedWoodruff, David J.; Sawyer, Richard L. – Applied Psychological Measurement, 1989
Two methods--non-distributional and normal--are derived for estimating measures of pass-fail reliability. Both are based on the Spearman Brown formula and require only a single test administration. Results from a simulation (n=20,000 examinees) and a licensure examination (n=4,828 examinees) illustrate these methods. (SLD)
Descriptors: Equations (Mathematics), Estimation (Mathematics), Licensing Examinations (Professions), Measures (Individuals)
Peer reviewedLevin, Joel R.; Subkoviak, Michael J. – Applied Psychological Measurement, 1977
Textbook calculations of statistical power or sample size follow from formulas that assume that the variables under consideration are measured without error. However, in the real world of behavioral research, errors of measurement cannot be neglected. The determination of sample size is discussed, and an example illustrates blocking strategy.…
Descriptors: Analysis of Covariance, Analysis of Variance, Error of Measurement, Hypothesis Testing
Peer reviewedRozeboom, William W. – Applied Psychological Measurement, 1989
Formulas are provided for estimating the reliability of a linear composite of non-equivalent subtests given the reliabilities of component subtests. The reliability of the composite is compared to that of its components. An empirical example uses data from 170 children aged 4 through 8 years performing 34 Piagetian tasks. (SLD)
Descriptors: Elementary School Students, Equations (Mathematics), Estimation (Mathematics), Mathematical Models
Peer reviewedEiting, Mindert H. – Applied Psychological Measurement, 1991
A method is proposed for sequential evaluation of reliability of psychometric instruments. Sample size is unfixed; a test statistic is computed after each person is sampled and a decision is made in each stage of the sampling process. Results from a series of Monte-Carlo experiments establish the method's efficiency. (SLD)
Descriptors: Computer Simulation, Equations (Mathematics), Estimation (Mathematics), Mathematical Models
Peer reviewedMenasco, Michael B.; Curry, David J. – Applied Psychological Measurement, 1978
Scores on the Role Construct Repertory Test exhibited significant correlations with other forms of cognitive functioning, including American College Test scores in science and mathematics for a group of 79 college students. The Grid Form of the test was used. Test-retest reliability was low. (Author/CTM)
Descriptors: Achievement Tests, Cognitive Processes, Cognitive Style, Cognitive Tests
Peer reviewedBarnes, Janet L.; Landy, Frank J. – Applied Psychological Measurement, 1979
Although behaviorally anchored rating scales have both intuitive and empirical appeal, they have not always yielded superior results in contrast with graphic rating scales. Results indicate that the choice of an anchoring procedure will depend on the nature of the actual rating process. (Author/JKS)
Descriptors: Behavior Rating Scales, Comparative Testing, Higher Education, Rating Scales


