Publication Date
| In 2026 | 0 |
| Since 2025 | 74 |
| Since 2022 (last 5 years) | 509 |
| Since 2017 (last 10 years) | 1084 |
| Since 2007 (last 20 years) | 2603 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 169 |
| Practitioners | 49 |
| Teachers | 32 |
| Administrators | 8 |
| Policymakers | 8 |
| Counselors | 4 |
| Students | 4 |
| Media Staff | 1 |
Location
| Turkey | 173 |
| Australia | 81 |
| Canada | 79 |
| China | 72 |
| United States | 56 |
| Taiwan | 44 |
| Germany | 43 |
| Japan | 41 |
| United Kingdom | 39 |
| Iran | 37 |
| Indonesia | 35 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 1 |
| Meets WWC Standards with or without Reservations | 1 |
| Does not meet standards | 1 |
Gonzales, Patrick; Calsyn, Christopher; Jocelyn, Leslie; Mak, Kitty; Kastberg, David; Arafeh, Sousan; Williams, Trevor; Tsen, Winnie – 2001
The Third International Mathematics and Science Study (TIMSS) is one of the most comprehensive international studies of schooling and students' achievement in science and mathematics. TIMSS was originally conducted in 1995. Four years later in 1999, the Third International Mathematics and Science Study-Repeat (TIMSS-R) was conducted. This document…
Descriptors: Academic Achievement, Criterion Referenced Tests, Foreign Countries, Grade 8
Peer reviewedBrennan, Robert L. – Journal of Educational Technology Systems, 1973
A review of four areas in which the computer has influenced the theory and practice of achievement testing: a) test scoring and item analysis, b) item sampling, c) item generation, and d) the sequencing of items resulting in various types of adaptive testing. Also, reference is made to the impact of computer assisted achievement testing upon the…
Descriptors: Achievement Tests, Adaptive Testing, Computer Assisted Instruction, Data Processing
Peer reviewedKolstad, Rosemarie K.; And Others – Journal of Research and Development in Education, 1983
A study compared college students' performance on complex multiple-choice tests with scores on multiple true-false clusters. Researchers concluded that the multiple-choice tests did not accurately measure students' knowledge and that cueing and guessing led to grade inflation. (PP)
Descriptors: Achievement Tests, Difficulty Level, Guessing (Tests), Higher Education
Peer reviewedMarsh, Herbert W. – Educational and Psychological Measurement, 1983
This study compares multitrait-multimethod analyses (MTMM) performed on items and factor scores derived from the items. Based upon this demonstration, researchers are encouraged to conduct a preliminary factor analysis before analyzing MTMM data whenever there is a doubt about the underlying trait structure. (Author/PN)
Descriptors: Factor Analysis, Higher Education, Item Analysis, Measurement Techniques
Peer reviewedSecolsky, Charles – Journal of Educational Measurement, 1983
A model is presented using examinee judgements in detecting ambiguous/misinterpreted items on teacher-made criterion-referenced tests. A computational example and guidelines for constructing domain categories and interpreting the indices are presented. (Author/PN)
Descriptors: Criterion Referenced Tests, Higher Education, Item Analysis, Mathematical Models
Peer reviewedBenbasat, Izak; And Others – Computers and Education, 1982
Describes a computer-assisted testing system which produces multiple-choice examinations for a college course in business administration. The system uses SPIRES (Stanford Public Information REtrieval System) to manage a database of questions and related data, mark-sense cards for machine grading tests, and ACL (6) (Audit Command Language) to…
Descriptors: Computer Assisted Testing, Data Processing, Databases, Higher Education
Peer reviewedSchmeiser, Cynthia B.; Ferguson, Richard L. – Educational Horizons, 1979
Test materials in English and social studies, reflecting Black and White culture, were developed and administered to Black and White high school students. Analyses of variance indicated that performance by either group was not affected by Black or White culture content. White examinees scored higher on all tests. (SJL)
Descriptors: Academic Achievement, Black Students, Content Analysis, Cultural Influences
Peer reviewedKirsch, Irwin S. – Journal of Educational Measurement, 1980
The construct validity of reading comprehension test items was studied in a two-stage process. Five characteristics of task difficulty were defined and a heterogeneous set of 52 items were rated for these characteristics. Then correlations were obtained between ratings and item difficulty data. (CTM)
Descriptors: Adults, Cognitive Processes, Difficulty Level, Evaluation Criteria
Peer reviewedDonlon, Thomas F.; And Others – Applied Psychological Measurement, 1980
The scope and nature of sex differences in the Graduate Record Examination are explored by identifying individual test items that differ from the other items in terms of the magnitude of the difference in item difficulty for the sexes. In general, limited evidence of differences was established. (Author/CTM)
Descriptors: Aptitude Tests, College Entrance Examinations, Graduate Students, Higher Education
Peer reviewedHansen, Chris J.; Zytowski, Donald G. – Educational and Psychological Measurement, 1979
A measure of the extent to which the Kuder Occupational Interest Survey (KOIS) was predictive of occupational membership for an individual was correlated with KOIS item and scale scores. Results indicated that the KOIS was a moderator of its own predictive validity. (Author/JKS)
Descriptors: Females, Followup Studies, Interest Inventories, Item Analysis
Green, Douglas W.; Hering, Jeffrey – Creative Computing, 1979
A computer program for analyzing test results is presented. Item analysis and the point biserial correlation coefficient, both generated by this program, are explained and discussed. (MK)
Descriptors: Computer Programs, Computers, Educational Testing, Evaluation
Peer reviewedGottfried, Gail M. – Journal of Child Language, 1997
Investigates comprehension of metaphoric compounds. The study asked children and adults to identify referents of various types of these compounds. Findings reveal that 5-year-olds outperformed 3-year-olds but performed significantly less well than adults and that 3-year-olds did not generally interpret the compounds literally. (37 references)…
Descriptors: Adults, Child Language, Cognitive Development, College Students
Peer reviewedKolstad, Rosemarie K.; Kolstad, Robert A. – Journal of Dental Education, 1991
A study evaluated the use of a "none-of-these" option on multiple-choice achievement tests in undergraduate dental education. Results indicated this option neither enhanced nor diminished examinee performance stability but did reduce the examinee's opportunity to select correct choices by means unrelated to course objectives, thereby enhancing…
Descriptors: Achievement Tests, Dental Schools, Difficulty Level, Higher Education
Peer reviewedJohnson, Sylvia T.; Wallace, Michael B. – Journal of Educational Measurement, 1989
The effects were assessed of a test coaching program for urban Black youth, who intended to take the Scholastic Aptitude Test (SAT), on their performance on quantitative items. Findings for 116 program participants are discussed in relation to the improvement of supplemental instructional programs. (TJH)
Descriptors: Black Students, College Entrance Examinations, High School Students, Item Analysis
Peer reviewedAiken, Lewis R. – Educational and Psychological Measurement, 1989
Two alternatives to traditional item analysis and reliability estimation procedures are considered for determining the difficulty, discrimination, and reliability of optional items on essay and other tests. A computer program to compute these measures is described, and illustrations are given. (SLD)
Descriptors: College Entrance Examinations, Computer Software, Difficulty Level, Essay Tests


