Publication Date
| In 2026 | 0 |
| Since 2025 | 197 |
| Since 2022 (last 5 years) | 1067 |
| Since 2017 (last 10 years) | 2577 |
| Since 2007 (last 20 years) | 4938 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 225 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 65 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Peer reviewedBensoussan, Marsha – System, 1982
Discusses the hypothesis that the difficulty of a multiple-choice comprehension test does not just depend on the difficulty of the text, but rather on the text and the questions. Presents ways to evaluate the difficulty of the text, the questions on it, and the test as a whole. (EKN)
Descriptors: English (Second Language), Language Tests, Multiple Choice Tests, Reading Comprehension
Peer reviewedVan der Linden, Wim J. – Journal of Educational Measurement, 1982
An ignored aspect of standard setting, namely the possibility that Angoff or Nedelsky judges specify inconsistent probabilities (e.g., low probabilities for easy items but large probabilities for hard items) is explored. A latent trait method is proposed to estimate such misspecifications, and an index of consistency is defined. (Author/PN)
Descriptors: Cutting Scores, Latent Trait Theory, Mastery Tests, Mathematical Models
Peer reviewedSchrock, Timothy J.; Mueller, Daniel J. – Journal of Educational Research, 1982
Three item-construction principles for multiple-choice tests were studied to determine how they affected test results for high school students: (1) use of incomplete sentence stem; (2) location of blank in the stem; and (3) presence of noncueing material. Differences in item construction had a slight effect on test results. (Authors/CJ)
Descriptors: Cues, High School Students, High Schools, Item Analysis
Peer reviewedAskegaard, Lewis D.; Umila, Benwardo V. – Journal of Educational Measurement, 1982
Multiple matrix sampling of items and examinees was applied to an 18-item rank order instrument administered to a randomly assigned group and compared to the ordering and ranking of all items by control subjects. High correlations between ranks suggest the methodology may viably reduce respondent effort on long rank ordering tasks. (Author/CM)
Descriptors: Evaluation Methods, Item Sampling, Junior High Schools, Student Reaction
Peer reviewedMishra, Shitala P. – Psychology in the Schools, 1982
Investigated cultural bias in 79 items of three verbal tests of the Wechsler Intelligence Scale for Children-Revised (WISC-R). Administered three subtests to 40 Anglo- and 40 Native-American Navajo subjects. Results indicated 15 of 79 items (information, similarities, and vocabulary subtests) were biased against the Navajo sample. (Author)
Descriptors: Anglo Americans, Children, Comparative Analysis, Culture Fair Tests
Stratton, N. J. – Teaching at a Distance, 1981
A study of recurrent faults in multiple-choice items in Britain's Open University's computer-marked tests has led to a procedure for avoiding these faults. A description of the study covers the incidence and sources of faults (obviousness, memorization, unclear instruction, ambiguity, distractors, inter-item effects, and structure) and…
Descriptors: Error Patterns, Foreign Countries, Higher Education, Item Analysis
Peer reviewedScheuneman, Janice – Journal of Educational Measurement, 1979
This paper presents a chi square method for assessing bias in test items. In this procedure an unbiased item is defined as an item for which the probability of a correct response is the same for any person of a given ability level regardless of that person's ethnic group. (Author/CTM)
Descriptors: Cultural Differences, Culture Fair Tests, Item Analysis, Primary Education
Peer reviewedPlake, Barbara S.; And Others – Journal of Experimental Education, 1981
Number right and elimination scores were analyzed on a college level mathematics exam assembled from pretest data. Anxiety measures were administered along with the experimental forms to undergraduates. Results suggest that neither test scores nor attitudes are influenced by item order knowledge thereof, or anxiety level. (Author/GK)
Descriptors: College Mathematics, Difficulty Level, Higher Education, Multiple Choice Tests
Imrie, B. W. – Assessment in Higher Education, 1979
Three case studies are presented that describe variations used to obtain student perceptions of examinations and examination questions. Student perceptions can provide feedback about the quality of the examination. Results are presented and discussed and some examples of test evaluation questions are given. (Author/MLW)
Descriptors: Case Studies, Educational Testing, Feedback, Higher Education
Peer reviewedRoid, Gale; Haladyna, Tom – Review of Educational Research, 1980
A continuum of item-writing methods is proposed ranging from informal-subjective methods to algorithmic-objective methods. Examples of techniques include objective-based item writing, amplified objectives, item forms, facet design, domain-referenced concept testing, and computerized techniques. (Author/CP)
Descriptors: Achievement Tests, Algorithms, Computer Assisted Testing, Criterion Referenced Tests
Peer reviewedForsyth, Robert A.; Spratt, Kevin F. – Journal of Educational Measurement, 1980
The effects of two item formats on item difficulty and item discrimination indices for mathematics problem solving multiple-choice tests were investigated. One format required identifying the proper "set-up" for the item; the other format required complete solving of the item. (Author/JKS)
Descriptors: Difficulty Level, Junior High Schools, Multiple Choice Tests, Problem Solving
Peer reviewedAlderman, Donald L.; Powers, Donald E. – American Educational Research Journal, 1980
Control groups' scores on the Scholastic Aptitude Test-Verbal were compared to scores of secondary school students who were exposed to their school's special test preparation programs. Special preparation resulted in an overall difference of eight points on the total score, corresponding to one additional correctly answered item. (Author/GDC)
Descriptors: College Bound Students, College Entrance Examinations, High Schools, Program Effectiveness
Peer reviewedPlake, Barbara S. – Educational and Psychological Measurement, 1980
Analysis of variance and subjective rating by curriculum specialists were used to identify biased items on the Iowa Tests of Basic Skills. Results show little agreement between statistical and subjective methods. Test developers should statistically support a reviewer's selection of biased items. (Author/CP)
Descriptors: Achievement Tests, Analysis of Variance, Elementary Education, Evaluation Methods
Peer reviewedMarr, Mary Beth; Lyon, Kathleen R. – Reading Psychology, 1981
Reveals that the "Classroom Reading Inventory" is the most passage independent of three informal reading inventories administered to fourth grade students. Shows that the general information, vocabulary meaning, and affective question categories tended to be particularly passage independent. (FL)
Descriptors: Content Analysis, Grade 4, Informal Reading Inventories, Intermediate Grades
Peer reviewedWhite, Edward M. – Journal of Basic Writing, 1978
Presents remarks on essay testing made by Edward M. White, an account of procedures for developing essay questions for the California State University and Colleges (CSUC) English Equivalency Examination, the 1974 essay scale with accompanying writing samples, and the questions and scale used in the 1975 test. (RL)
Descriptors: Equivalency Tests, Essay Tests, Higher Education, Models


