NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20260
Since 20250
Since 2022 (last 5 years)0
Since 2017 (last 10 years)2
Since 2007 (last 20 years)3
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 25 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip – Journal of Educational and Behavioral Statistics, 2017
An increasing concern of producers of educational assessments is fraudulent behavior during the assessment (van der Linden, 2009). Benefiting from item preknowledge (e.g., Eckerly, 2017; McLeod, Lewis, & Thissen, 2003) is one type of fraudulent behavior. This article suggests two new test statistics for detecting individuals who may have…
Descriptors: Test Items, Cheating, Testing Problems, Identification
Peer reviewed Peer reviewed
Direct linkDirect link
Debeer, Dries; Janssen, Rianne; De Boeck, Paul – Journal of Educational Measurement, 2017
When dealing with missing responses, two types of omissions can be discerned: items can be skipped or not reached by the test taker. When the occurrence of these omissions is related to the proficiency process the missingness is nonignorable. The purpose of this article is to present a tree-based IRT framework for modeling responses and omissions…
Descriptors: Item Response Theory, Test Items, Responses, Testing Problems
Peer reviewed Peer reviewed
Direct linkDirect link
Taskinen, Päivi H.; Steimel, Jochen; Gräfe, Linda; Engell, Sebastian; Frey, Andreas – Peabody Journal of Education, 2015
This study examined students' competencies in engineering education at the university level. First, we developed a competency model in one specific field of engineering: process dynamics and control. Then, the theoretical model was used as a frame to construct test items to measure students' competencies comprehensively. In the empirical…
Descriptors: Models, Engineering Education, Test Items, Outcome Measures
Peer reviewed Peer reviewed
Cziko, Gary A. – Educational and Psychological Measurement, 1984
Some problems associated with the criteria of reproducibility and scalability as they are used in Guttman scalogram analysis to evaluate cumulative, nonparametric scales of dichotomous items are discussed. A computer program is presented which analyzes response patterns elicited by dichotomous scales designed to be cumulative. (Author/DWH)
Descriptors: Scaling, Statistical Analysis, Test Construction, Test Items
PDF pending restoration PDF pending restoration
Kogut, Jan – 1986
Methods and indices based on item response theory (IRT) for detecting and diagnosing aberrant response patterns are reviewed. These indices are divided into three groups: (1) residuals-based; (2) likelihood-based; and (3) ratio of covariances-based (extended cautions). For each index, the determination of its sampling distribution as well as its…
Descriptors: Algorithms, Computer Simulation, Foreign Countries, Latent Trait Theory
McArthur, David L. – 1981
Item bias, when present in a multiple-choice test, can be detected by appropriate analyses of the persons x items scoring matrix. Five related schemes for the statistical analysis of bias were applied to a widely used, primary skills multiple-choice test which was administered in either its English- or Spanish-language version at each of the two…
Descriptors: Comparative Analysis, Elementary Education, Multiple Choice Tests, Spanish
Peer reviewed Peer reviewed
Frary, Robert B.; Hutchinson, T.P. – Educational and Psychological Measurement, 1982
Alternate versions of Hutchinson's theory were compared, and one which implies the existence of partial knowledge was found to be better than one which implies that an appropriate measure of ability is obtained by applying the conventional correction for guessing. (Author/PN)
Descriptors: Guessing (Tests), Latent Trait Theory, Multiple Choice Tests, Scoring Formulas
Wilcox, Rand R. – 1978
Two fundamental problems in mental test theory are to estimate true score and to estimate the amount of error when testing an examinee. In this report, three probability models which characterize a single test item in terms of a population of examinees are described. How these models may be modified to characterize a single examinee in terms of an…
Descriptors: Achievement Tests, Comparative Analysis, Error of Measurement, Mathematical Models
Merz, William R. – 1980
Several methods of assessing test item bias are described, and the concept of fair use of tests is examined. A test item is biased if individuals of equal ability have different probabilities of attaining the item correct. The following seven general procedures used to examine test items for bias are summarized and discussed: (1) analysis of…
Descriptors: Comparative Analysis, Evaluation Methods, Factor Analysis, Mathematical Models
Kelderman, Henk – 1986
A method is proposed for the detection of item bias with respect to observed or unobserved subgroups. The method uses quasi-loglinear models for the incomplete subgroup x test score x item 1 x ... x item k contingency table. If the subgroup membership is unknown, the models are the incomplete-latent-class models of S. J. Haberman (1979). The…
Descriptors: Foreign Countries, Higher Education, Latent Trait Theory, Mathematical Models
Kelderman, Henk; Macready, George B. – 1988
The use of loglinear latent class models to detect item bias was studied. Purposes of the study were to: (1) develop procedures for use in assessing item bias when the grouping variable with respect to which bias occurs is not observed; (2) develop bias detection procedures that relate to a conceptually different assessed trait--a categorical…
Descriptors: Foreign Countries, Higher Education, Latent Trait Theory, Mathematical Models
Kuntz, Patricia – 1982
The quality of mathematics multiple choice items and their susceptibility to test wiseness were examined. Test wiseness was defined as "a subject's capacity to utilize the characteristics and formats of the test and/or test taking situation to receive a high score." The study used results of the Graduate Record Examinations Aptitude Test (GRE) and…
Descriptors: Cues, Item Analysis, Multiple Choice Tests, Psychometrics
Livingston, Samuel A. – 1986
This paper deals with test fairness regarding a test consisting of two parts: (1) a "common" section, taken by all students; and (2) a "variable" section, in which some students may answer a different set of questions from other students. For example, a test taken by several thousand students each year contains a common multiple-choice portion and…
Descriptors: Difficulty Level, Error of Measurement, Essay Tests, Mathematical Models
Peer reviewed Peer reviewed
Gohmann, Stephan F.; Spector, Lee C. – Journal of Economic Education, 1989
Compares the effect of content ordering and scrambled ordering on examinations in courses, such as economics, that require quantitative skills. Empirical results suggest that students do no better if they are given a content-ordered rather than a scrambled examination as student performance is not adversely affected by scrambled ordered…
Descriptors: Cheating, Economics Education, Educational Research, Grading
Theunissen, Phiel J. J. M. – 1983
Any systematic approach to the assessment of students' ability implies the use of a model. The more explicit the model is, the more its users know about what they are doing and what the consequences are. The Rasch model is a strong model where measurement is a bonus of the model itself. It is based on four ideas: (1) separation of observable…
Descriptors: Ability Grouping, Difficulty Level, Evaluation Criteria, Item Sampling
Previous Page | Next Page »
Pages: 1  |  2