Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 0 |
| Since 2007 (last 20 years) | 1 |
Descriptor
| Comparative Testing | 18 |
| Computer Assisted Testing | 7 |
| Higher Education | 7 |
| Mathematical Models | 6 |
| Test Format | 6 |
| Adaptive Testing | 5 |
| Item Response Theory | 5 |
| Test Items | 5 |
| Foreign Countries | 4 |
| Multiple Choice Tests | 4 |
| Scoring | 4 |
| More ▼ | |
Source
| Applied Psychological… | 18 |
Author
Publication Type
| Journal Articles | 16 |
| Reports - Evaluative | 8 |
| Reports - Research | 8 |
| Speeches/Meeting Papers | 1 |
| Tests/Questionnaires | 1 |
Education Level
Audience
Laws, Policies, & Programs
Assessments and Surveys
| ACT Assessment | 1 |
| Advanced Placement… | 1 |
| Armed Services Vocational… | 1 |
| Differential Aptitude Test | 1 |
What Works Clearinghouse Rating
Kluge, Annette – Applied Psychological Measurement, 2008
The use of microworlds (MWs), or complex dynamic systems, in educational testing and personnel selection is hampered by systematic measurement errors because these new and innovative item formats are not adequately controlled for their difficulty. This empirical study introduces a way to operationalize an MW's difficulty and demonstrates the…
Descriptors: Personnel Selection, Self Efficacy, Educational Testing, Computer Uses in Education
Peer reviewedStocking, Martha L.; And Others – Applied Psychological Measurement, 1993
A method of automatically selecting items for inclusion in a test with constraints on item content and statistical properties was applied to real data. Tests constructed manually from the same data and constraints were compared to tests constructed automatically. Results show areas in which automated assembly can improve test construction. (SLD)
Descriptors: Algorithms, Automation, Comparative Testing, Computer Assisted Testing
Peer reviewedDrasgow, Fritz; And Others – Applied Psychological Measurement, 1991
Extensions of unidimensional appropriateness indices are developed for multiunidimensional tests (multidimensional tests composed of unidimensional subtests). Simulated and real data (scores of 2,978 students on the Armed Services Vocational Aptitude Battery) were used to evaluate the indices' effectiveness in determining individuals who are…
Descriptors: Comparative Testing, Computer Simulation, Equations (Mathematics), Graphs
Peer reviewedBetz, Nancy E. – Applied Psychological Measurement, 1977
The effects of providing immediate feedback on an item-by-item basis to testees were investigated on several samples of undergraduates who were administered either adaptive or conventional tests. Results generally indicated that knowledge of results improved test performance. (JKS)
Descriptors: Academic Aptitude, Adaptive Testing, Comparative Testing, Feedback
Peer reviewedHenly, Susan J.; And Others – Applied Psychological Measurement, 1989
A group of covariance structure models was examined to ascertain the similarity between conventionally administered and computerized adaptive versions of the Differential Aptitude Test (DAT). Results for 332 students indicate that the computerized version of the DAT is an adequate representation of the conventional test battery. (TJH)
Descriptors: Ability Identification, Adaptive Testing, Comparative Testing, Computer Assisted Testing
Peer reviewedReilly, Richard R.; Echternacht, Gary J. – Applied Psychological Measurement, 1979
Occupational interest scales and homogeneous scales were developed and compared on a sample of 3,072 Air Force personnel in a double cross-validation design. Results indicated that the occupational scales were slightly superior in correctly classifying satisified personnel v people in general. (Author/JKS)
Descriptors: Attitude Measures, Classification, Comparative Testing, Interest Inventories
Peer reviewedLunz, Mary E.; And Others – Applied Psychological Measurement, 1992
The effects of reviewing items and altering responses on the efficiency of computerized adaptive tests and resultant ability estimates of the examinees were explored for medical technology students (220 students could and 492 students could not review and alter their responses). Data do not support disallowing review. (SLD)
Descriptors: Ability, Adaptive Testing, Comparative Testing, Computer Assisted Testing
Peer reviewedNorcini, John; And Others – Applied Psychological Measurement, 1991
Effects of numbers of experts (NOEs) and common items (CIs) on the scaling of cutting scores from expert judgments were studied for 11,917 physicians taking 2 forms of a medical specialty examination. Increasing NOEs and CIs reduced error; beyond 5 experts and 25 CIs, error differences were small. (SLD)
Descriptors: Comparative Testing, Cutting Scores, Equated Scores, Estimation (Mathematics)
Peer reviewedBennett, Randy Elliot; And Others – Applied Psychological Measurement, 1990
The relationship of an expert-system-scored constrained free-response item type to multiple-choice and free-response items was studied using data for 614 students on the College Board's Advanced Placement Computer Science (APCS) Examination. Implications for testing and the APCS test are discussed. (SLD)
Descriptors: College Students, Comparative Testing, Computer Assisted Testing, Computer Science
Peer reviewedTraub, Ross E.; Fisher, Charles W. – Applied Psychological Measurement, 1977
Two sets of mathematical reasoning and two sets of verbal comprehension items were cast into each of three formats--constructed response, standard multiple-choice, and Coombs multiple-choice--in order to assess whether tests with identical content but different formats measure the same attribute. (Author/CTM)
Descriptors: Comparative Testing, Confidence Testing, Constructed Response, Factor Analysis
Peer reviewedBarnes, Janet L.; Landy, Frank J. – Applied Psychological Measurement, 1979
Although behaviorally anchored rating scales have both intuitive and empirical appeal, they have not always yielded superior results in contrast with graphic rating scales. Results indicate that the choice of an anchoring procedure will depend on the nature of the actual rating process. (Author/JKS)
Descriptors: Behavior Rating Scales, Comparative Testing, Higher Education, Rating Scales
Peer reviewedDe Ayala, R. J. – Applied Psychological Measurement, 1992
A computerized adaptive test (CAT) based on the nominal response model (NR CAT) was implemented, and the performance of the NR CAT and a CAT based on the three-parameter logistic model was compared. The NR CAT produced trait estimates comparable to those of the three-parameter test. (SLD)
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Equations (Mathematics)
Peer reviewedKim, Seock-Ho; Cohen, Allan S. – Applied Psychological Measurement, 1991
The exact and closed-interval area measures for detecting differential item functioning are compared for actual data from 1,000 African-American and 1,000 white college students taking a vocabulary test with items intentionally constructed to favor 1 set of examinees. No real differences in detection of biased items were found. (SLD)
Descriptors: Black Students, College Students, Comparative Testing, Equations (Mathematics)
van der Linden, Wim J. – Applied Psychological Measurement, 2006
Two local methods for observed-score equating are applied to the problem of equating an adaptive test to a linear test. In an empirical study, the methods were evaluated against a method based on the test characteristic function (TCF) of the linear test and traditional equipercentile equating applied to the ability estimates on the adaptive test…
Descriptors: Adaptive Testing, Computer Assisted Testing, Test Format, Equated Scores
Peer reviewedBirenbaum, Menucha; And Others – Applied Psychological Measurement, 1992
The effect of multiple-choice (MC) or open-ended (OE) response format on diagnostic assessment of algebra test performance was investigated with 231 eighth and ninth graders in Tel Aviv (Israel) using bug or rule space analysis. Both analyses indicated closer similarity between parallel OE subsets than between stem-equivalent OE and MC subsets.…
Descriptors: Algebra, Comparative Testing, Educational Assessment, Educational Diagnosis
Previous Page | Next Page ยป
Pages: 1 | 2
Direct link
