Publication Date
| In 2026 | 0 |
| Since 2025 | 197 |
| Since 2022 (last 5 years) | 1067 |
| Since 2017 (last 10 years) | 2577 |
| Since 2007 (last 20 years) | 4938 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 225 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 65 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Pang, Xiao L.; And Others – 1994
The function of Mantel-Haenszel (MH) and logistic regression (LR) statistics with real data in detecting gender-based differentially functioning items (DIF) was investigated when sample size and criterion variable varied. The data base consisted of the item responses of a population of 183,356 Caucasians to the Math test of the ACT Assessment…
Descriptors: College Entrance Examinations, Foreign Countries, Identification, Item Bias
Bergstrom, Betty; And Others – 1994
Examinee response times from a computerized adaptive test taken by 204 examinees taking a certification examination were analyzed using a hierarchical linear model. Two equations were posed: a within-person model and a between-person model. Variance within persons was eight times greater than variance between persons. Several variables…
Descriptors: Adaptive Testing, Adults, Certification, Computer Assisted Testing
PDF pending restorationSykes, Robert C.; And Others – 1996
The presence of multiple readings of a student response to a constructed-response item in a large-scale assessment requires a procedure for combining the ratings to obtain an item score. An alternative to the averaged item ratings that are usually used is the summing of ratings for each item. This study evaluated the effect of summing as opposed…
Descriptors: Constructed Response, High Schools, Item Response Theory, Mathematics Education
Kehoe, Jerard – 1995
This digest describes some basics of the construction of multiple-choice tests. As a rule, the test maker should strive for test item stems (introductory questions or incomplete statements at the beginning of each item that are followed by the options) that are clear and parsimonious, answers that are unequivocal and chosen by the students who do…
Descriptors: Culture Fair Tests, Distractors (Tests), Educational Assessment, Item Bias
Shorey, Leonard – 1991
Tests in social studies and integrated science given in Saint Vincent, Saint Lucia, Grenada, and Dominica were analyzed by the Organization for Co-operation in Overseas Development (OCOD) Comprehensive Teacher Training Program (CTTP) for discrimination, difficulty, and reliability, as well as other characteristics. There were 767 examinees for the…
Descriptors: Difficulty Level, Elementary Secondary Education, Evaluation Methods, Foreign Countries
Burstein, Jill C.; Kaplan, Randy M. – 1995
There is a considerable interest at Educational Testing Service (ETS) to include performance-based, natural language constructed-response items on standardized tests. Such items can be developed, but the projected time and costs required to have these items scored by human graders would be prohibitive. In order for ETS to include these types of…
Descriptors: Computer Assisted Testing, Constructed Response, Cost Effectiveness, Hypothesis Testing
Schnipke, Deborah L.; Pashley, Peter J. – 1997
Differences in test performance on time-limited tests may be due in part to differential response-time rates between subgroups, rather than real differences in the knowledge, skills, or developed abilities of interest. With computer-administered tests, response times are available and may be used to address this issue. This study investigates…
Descriptors: Computer Assisted Testing, Data Analysis, English, High Stakes Tests
Alberta Dept. of Education, Edmonton. Student Evaluation Branch. – 1997
Intended for students taking the Grade 12 Diploma Examinations in English 30, this "questions booklet" presents 70 multiple choice test items based on 8 reading selections in the accompanying readings booklet. After instructions for students, the booklet presents the multiple choice items which test students' comprehension of the poetry,…
Descriptors: Achievement Tests, Drama, English Instruction, Foreign Countries
Muthen, Bengt; And Others – 1995
A set of methods is proposed for the analysis of opportunity to learn (OTL) in relation to achievement in large-scale educational assessments. The focus is on how to assess the effect of OTL on performance while taking prior performance and other background factors into account. The methods are illustrated with mathematics data from the National…
Descriptors: Academic Achievement, Achievement Tests, Educational Assessment, Mathematics Tests
De Ayala, R. J. – 1992
One important and promising application of item response theory (IRT) is computerized adaptive testing (CAT). The implementation of a nominal response model-based CAT (NRCAT) was studied. Item pool characteristics for the NRCAT as well as the comparative performance of the NRCAT and a CAT based on the three-parameter logistic (3PL) model were…
Descriptors: Adaptive Testing, Comparative Testing, Computer Assisted Testing, Computer Simulation
Herd, Amon – 1994
This notebook contains criterion-referenced test items for testing students' knowledge of small engines. The test items are based upon competencies found in the Missouri Small Engine Competency Profile. The test item bank is organized in 18 sections that cover the following duties: shop procedures; tools and equipment; fasteners; servicing fuel…
Descriptors: Auto Mechanics, Competence, Competency Based Education, Criterion Referenced Tests
Wainer, Howard; Thissen, David – 1994
When an examination consists in whole or part of constructed response test items, it is common practice to allow the examinee to choose a subset of the constructed response questions from a larger pool. It is sometimes argued that, if choice were not allowed, the limitations on domain coverage forced by the small number of items might unfairly…
Descriptors: Constructed Response, Difficulty Level, Educational Testing, Equated Scores
Tannenbaum, Richard J. – 1994
A job analysis was conducted, focusing on the knowledge and abilities important for beginning French teachers. The results of the job analysis are to be used to define the content domain of the subject assessment in French for the Praxis series of professional assessments for beginning teachers. A domain of 212 knowledge statements and ability…
Descriptors: Administrators, Beginning Teachers, Cutting Scores, Educational Assessment
Powell, Z. Emily – 1992
Little research exists on the psychological impacts of computerized adaptive testing (CAT) and how it may affect test performance. Three CAT procedures were examined, in which items were selected to match students' achievement levels, from the item pool at random, or according to student choice of item difficulty levels. Twenty-four graduate…
Descriptors: Academic Achievement, Adaptive Testing, Comparative Testing, Computer Assisted Testing
Ackerman, Terry A.; Evans, John A. – 1992
The relationship between levels of reliability and the power of two bias and differential item functioning (DIF) detection methods is examined. Both methods, the Mantel-Haenszel (MH) procedure of P. W. Holland and D. T. Thayer (1988) and the Simultaneous Item Bias (SIB) procedure of R. Shealy and W. Stout (1991), use examinees' raw scores as a…
Descriptors: Comparative Analysis, Equations (Mathematics), Error of Measurement, Item Bias


