Publication Date
| In 2026 | 0 |
| Since 2025 | 197 |
| Since 2022 (last 5 years) | 1067 |
| Since 2017 (last 10 years) | 2577 |
| Since 2007 (last 20 years) | 4938 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 653 |
| Teachers | 563 |
| Researchers | 250 |
| Students | 201 |
| Administrators | 81 |
| Policymakers | 22 |
| Parents | 17 |
| Counselors | 8 |
| Community | 7 |
| Support Staff | 3 |
| Media Staff | 1 |
| More ▼ | |
Location
| Turkey | 225 |
| Canada | 223 |
| Australia | 155 |
| Germany | 116 |
| United States | 99 |
| China | 90 |
| Florida | 86 |
| Indonesia | 82 |
| Taiwan | 78 |
| United Kingdom | 73 |
| California | 65 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 4 |
| Meets WWC Standards with or without Reservations | 4 |
| Does not meet standards | 1 |
Peer reviewedWise, Steven L.; Finney, Sara J.; Enders, Craig K.; Freeman, Sharon A.; Severance, Donald D. – Applied Measurement in Education, 1999
Examined whether providing item review on a computerized adaptive test could be used by examinees to inflate their scores. Two studies involving 139 undergraduates suggest that examinees are not highly proficient at discriminating item difficulty. A simulation study showed the usefulness of a strategy identified by G. Kingsbury (1996) as a way to…
Descriptors: Adaptive Testing, Computer Assisted Testing, Difficulty Level, Higher Education
Peer reviewedMooney, G. A.; Bligh, J. G.; Leinster, S. J. – Medical Teacher, 1998
Presents a system of classification for describing computer-based assessment techniques based on the level of action and educational activity they offer. Illustrates 10 computer-based assessment techniques and discusses their educational value. Contains 14 references. (Author)
Descriptors: Adaptive Testing, Classification, Computer Assisted Testing, Foreign Countries
Peer reviewedGasser, Michael B.; Tan, Rowena N. – Educational and Psychological Measurement, 1999
Developed a scale of cultural tolerance, examined the latent structure of the scale, and determined its relationship to other relevant constructs. Results with samples of 262, 165, and 86 college students suggest a two-dimensional structure of items assessing affect and belief and items assessing intention to behave. Contains 41 references. (SLD)
Descriptors: Affective Behavior, Beliefs, College Students, Cultural Awareness
Peer reviewedVispoel, Walter P. – Journal of Educational Measurement, 1998
Studied effects of administration mode [computer adaptive test (CAT) versus self-adaptive test (SAT)], item-by-item answer feedback, and test anxiety on results from computerized vocabulary tests taken by 293 college students. CATs were more reliable than SATs, and administration time was less when feedback was provided. (SLD)
Descriptors: Adaptive Testing, College Students, Computer Assisted Testing, Feedback
Peer reviewedHamilton, Laura S. – Applied Measurement in Education, 1999
Explored methods for detecting gender-based differential item functioning on a science test that was part of the National Education Longitudinal Study of 1988. Exploration of conditioning variables and interviews with 25 high school students show a large male advantage for one item in particular. Discusses the basis of gender differences. (SLD)
Descriptors: Constructed Response, High School Students, High Schools, Interviews
Peer reviewedQuinn, Robert J.; Obenchain, Kathryn M. – Clearing House, 1999
Describes how students in a general secondary methods course responded to a gender-neutral exam question by consistently assuming that the student in the hypothetical scenario was male. Describes the follow-up class discussion, noting students' assumptions, defensive responses, subconscious decision making, and awareness/nonawareness of their own…
Descriptors: Higher Education, Methods Courses, Secondary Education, Sex Bias
Peer reviewedTakala, Sauli; Kaftandjieva, Felianka – Language Testing, 2000
Analyzes gender-uniform differential item functioning (DIF) in a second language vocabulary test with the tools of item response theory to study potential gender impact on the test performance measured by different item composites. Results show that while there are test items with indications of DIF in favor of either females or males, the test as…
Descriptors: English (Second Language), Foreign Countries, Item Analysis, Language Tests
Peer reviewedMount, Robert E.; Schumacker, Randall E. – Journal of Outcome Measurement, 1998
A Monte Carlo study was conducted using simulated dichotomous data to determine the effects of guessing on Rasch item fit statistics and the Logit Residual Index. Results indicate that no significant differences were found between the mean Rasch item fit statistics for each distribution type as the probability of guessing the correct answer…
Descriptors: Goodness of Fit, Guessing (Tests), Item Response Theory, Monte Carlo Methods
Peer reviewedScheuneman, Janice Dowd; Subhiyah, Raja G. – Journal of Outcome Measurement, 1998
An analysis of differential item functioning (DIF) in a medical specialty certification examination taken by 416 candidates through a simple Rasch model is presented. These analysis results serve as validity evidence for use of the Rasch model as a way to examine DIF in examinations. It is concluded that complex procedures are not necessary to…
Descriptors: Certification, Identification, Item Bias, Item Response Theory
Peer reviewedMislevy, Robert J. – Applied Measurement in Education, 1998
How summarizing National Assessment of Educational Progress (NAEP) results in terms of a market basket of items would affect achievement-level reporting is discussed. A market basket is a specific set of items one may administer, scores on which constitute a reporting scale. Advantages and limitations of the market-basket approach are discussed.…
Descriptors: Academic Achievement, Achievement Tests, Elementary Secondary Education, Item Banks
Peer reviewedBielinski, John; Davison, Mark L. – Journal of Educational Measurement, 2001
Used mathematics achievement data from the 1992 National Assessment of Educational Progress, the Third International Mathematics and Science Study, and the National Education Longitudinal Study of 1988 to examine the sex difference by item difficulty interaction. The predicted negative correlation was found for all eight populations and was…
Descriptors: Correlation, Difficulty Level, Interaction, Mathematics Tests
Peer reviewedDavidson, Fred – System, 2000
Statistical analysis tools in language testing are described, chiefly classical test theory and item response theory. Computer software for statistical analysis is briefly reviewed and divided into three tiers: commonly available; statistical packages; and specialty software. (Author/VWL)
Descriptors: Computer Software, Language Tests, Second Language Learning, Statistical Analysis
Peer reviewedEllis, Barbara B.; Mead, Alan D. – Educational and Psychological Measurement, 2000
Used the differential functioning of items and tests (DFIT) framework to examine the measurement equivalence of a Spanish translation of the Sixteen Personality Factor (16PF) Questionnaire using samples of 309 Anglo American college students and other adults, 280 English-speaking Hispanics, and 244 Spanish-speaking college students. Results show…
Descriptors: Adults, College Students, Higher Education, Hispanic American Students
Peer reviewedWang, Wen-Chung – Journal of Applied Measurement, 2000
Proposes a factorial procedure for investigating differential distractor functioning in multiple choice items that models each distractor with a distinct distractibility parameter. Results of a simulation study show that the parameters of the proposed modeling were recovered very well. Analysis of 10 4-choice items from a college entrance…
Descriptors: College Entrance Examinations, Distractors (Tests), Factor Structure, Foreign Countries
Peer reviewedReise, Steven P. – Applied Psychological Measurement, 2001
This book contains a series of research articles about computerized adaptive testing (CAT) written for advanced psychometricians. The book is divided into sections on: (1) item selection and examinee scoring in CAT; (2) examples of CAT applications; (3) item banks; (4) determining model fit; and (5) using testlets in CAT. (SLD)
Descriptors: Adaptive Testing, Computer Assisted Testing, Goodness of Fit, Item Banks


