Publication Date
| In 2026 | 0 |
| Since 2025 | 62 |
| Since 2022 (last 5 years) | 388 |
| Since 2017 (last 10 years) | 831 |
| Since 2007 (last 20 years) | 1345 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 195 |
| Teachers | 161 |
| Researchers | 93 |
| Administrators | 50 |
| Students | 34 |
| Policymakers | 15 |
| Parents | 12 |
| Counselors | 2 |
| Community | 1 |
| Media Staff | 1 |
| Support Staff | 1 |
| More ▼ | |
Location
| Canada | 63 |
| Turkey | 59 |
| Germany | 41 |
| United Kingdom | 37 |
| Australia | 36 |
| Japan | 35 |
| China | 33 |
| United States | 32 |
| California | 25 |
| Iran | 25 |
| United Kingdom (England) | 25 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Rentz, R. Robert – New Directions for Testing and Measurement, 1981
In this speculative article, Robert R. Rentz predicts that state educational assessment programs will continue to be fertile ground for measurement innovations, both technological and otherwise, in the areas of test content, development, administration and the reporting of test results. (AEF)
Descriptors: Educational Innovation, Prediction, State Programs, Test Construction
Peer reviewedLubin, Bernard; And Others – Journal of Clinical Psychology, 1981
Studied potential checking and response bias on the Depression Adjective Check Lists (DACL) by comparing scores of college students on true-false, forced-choice, and standard formats. By demonstrating only a weak social desirability response bias and failing to reveal checking bias, results support using the standard format. (Author)
Descriptors: Comparative Analysis, Depression (Psychology), Psychometrics, Response Style (Tests)
Peer reviewedBaker, Frank B. – Applied Psychological Measurement, 1996
Using the characteristic curve method for dichotomously scored test items, the sampling distributions of equating coefficients were examined. Simulations indicate that for the equating conditions studied, the sampling distributions of the equating coefficients appear to have acceptable characteristics, suggesting confidence in the values obtained…
Descriptors: Equated Scores, Item Response Theory, Sampling, Statistical Distributions
Haynie, W. J., III – Journal of Technology Education, 2003
Undergraduates were divided into four groups: (1) no study questions, multiple-choice test (n=71); (2) no questions, take-home test (n=71); (3) questions, no test (n=70); and (4) no questions, no test (n=67). The take-home group scored worst of all on novel information. They seem to have focused only on the required answers. (Contains 25…
Descriptors: Higher Education, Retention (Psychology), Study Guides, Study Habits
Peer reviewedCampbell, Todd; And Others – Educational and Psychological Measurement, 1997
The construct validity of scores from the Bem Sex-Role Inventory was studied using confirmatory factor analysis methods on data from 791 subjects. Measurement characteristics of the long and short forms were studied, with the short form yielding more reliable scores, as has previously been indicated. (Author/SLD)
Descriptors: Adults, Construct Validity, Factor Structure, Scores
Peer reviewedHarland, Lynn K. – Human Resource Development Quarterly, 2003
Leadership development students were randomly assigned to complete a personality questionnaire either as a normative test (n=85), forced-choice (n=84), or forced-choice with explanation of its benefits (n=86). Forced-choice-only takers perceived it as less accurate, useful, and respectful. Explanations influenced only perceptions of control, not…
Descriptors: Feedback, Forced Choice Technique, Leadership Training, Negative Attitudes
Shick, Jacqueline – Health Education (Washington D.C.), 1989
This article focuses on common errors associated with true-false, matching, completion, and essay questions as presented in textbook test manuals. Teachers should be able to select and/or adapt test questions which would be applicable to the content of their courses and which meet minimal standards for test construction. (JD)
Descriptors: Health Education, Higher Education, Secondary Education, Test Construction
Peer reviewedRoyer, James M.; And Others – Journal of Educational Research, 1989
This study evaluated multiple-choice and sentence verification technique items as on-line measures of comprehension while subjects (N=940 Air Force enlistees) read computer-presented text. Also evaluated was whether presentation of on-line questions had a positive impact on text learning. (IAH)
Descriptors: Adults, Computer Assisted Instruction, Enlisted Personnel, Reading Comprehension
Peer reviewedPlake, Barbara S.; And Others – Educational and Psychological Measurement, 1988
The effect of item context on differential item performance based on gender on mathematics test items was studied, using 404 male and 375 female adults. The analyses were based on a modified one-parameter item response theory methodology. Gender differences emerged; however, they may be due to chance. (TJH)
Descriptors: Achievement Tests, Adults, Latent Trait Theory, Mathematics Tests
Stape, Christopher J. – Performance and Instruction, 1995
Suggests methods for developing higher level objective test questions. Taxonomies that define learning outcomes are discussed; and examples for various test types are presented, including multiple correct answers; more complex forms, including classification and multiple true-false; relations and correlates; and interpretive exercises. (LRW)
Descriptors: Classification, Objective Tests, Outcomes of Education, Test Construction
Peer reviewedGuttman, Louis; Levy, Shlomit – Intelligence, 1991
Two structural laws for intelligence tests are discussed: one law concerns the sign of correlation coefficients and gives conditions under which all correlations between test items will be positive; and one law concerns the relative sizes of the correlation coefficients between intelligence items. A cylindrical structure extends these laws. (SLD)
Descriptors: Correlation, Foreign Countries, Intelligence Tests, Test Construction
Peer reviewedDorans, Neil J.; Lawrence, Ida M. – Applied Measurement in Education, 1990
A procedure for checking the score equivalence of nearly identical editions of a test is described and illustrated with Scholastic Aptitude Test data. The procedure uses the standard error of equating and uses graphical representation of score conversion deviations from the identity function in standard error units. (SLD)
Descriptors: Equated Scores, Grade Equivalent Scores, Scores, Statistical Analysis
Peer reviewedStern, Paul C.; Guagnano, Gregory A.; Dietz, Thomas – Educational and Psychological Measurement, 1998
A brief version of the instrument developed by S. Schwartz (1992, 1994) to measure the structure and content of human values was developed. Studies with 199 adults and 420 adults support the reliability of scores produced by the brief inventory's four three-item scales. Uses of the brief form are discussed. (SLD)
Descriptors: Adults, Reliability, Scores, Test Construction
Peer reviewedDemsky, Yvonne; Gass, Carlton; Edwards, William T.; Golden, Charles J. – Assessment, 1998
Investigated optimal two-, three-, four-, and five-test short forms of the Escala de Inteligencia Wechsler para Adultos (EIWA), the Spanish form of the Wechsler Adult Intelligence Scale (D. Wechsler, 1956). Results with 616 adults suggest that use of the EIWA should be limited to research and tracking cognitive changes over time. (SLD)
Descriptors: Adults, Culture Fair Tests, Intelligence Tests, Norms
Peer reviewedBoles, Larry – Journal of Communication Disorders, 1998
Gains made by a woman with Broca's aphasia as documented by traditional measures were paralleled by changes in conversation, including increased verbal output and efficiency, and changes in conversation-repair patterns. The progress documented with conversational discourse analysis was not observable from other test measures. (Author/DB)
Descriptors: Adults, Aphasia, Case Studies, Evaluation Methods


