Descriptor
| Mathematical Models | 13 |
| Testing Problems | 13 |
| Item Analysis | 4 |
| Test Validity | 4 |
| Cutting Scores | 3 |
| Equated Scores | 3 |
| Error of Measurement | 3 |
| Sampling | 3 |
| Scoring | 3 |
| Test Interpretation | 3 |
| Test Items | 3 |
| More ▼ | |
Source
| Journal of Educational… | 13 |
Author
| Wainer, Howard | 2 |
| Al-Karni, Ali | 1 |
| Baker, Frank B. | 1 |
| Beuk, Cees H. | 1 |
| Budescu, David | 1 |
| Foreman, Dale I. | 1 |
| Harris, Chester W. | 1 |
| Koffler, Stephen L. | 1 |
| Lewis, Charles | 1 |
| Roberts, Dennis M. | 1 |
| Secolsky, Charles | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 10 |
| Reports - Research | 7 |
| Reports - Evaluative | 3 |
Education Level
Audience
| Researchers | 1 |
Location
| Netherlands | 1 |
| New Jersey | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| SAT (College Admission Test) | 2 |
What Works Clearinghouse Rating
Peer reviewedWilcox, Rand R.; Harris, Chester W. – Journal of Educational Measurement, 1977
Emrick's proposed method for determining a mastery level cut-off score is questioned. Emrick's method is shown to be useful only in limited situations. (JKS)
Descriptors: Correlation, Cutting Scores, Mastery Tests, Mathematical Models
Peer reviewedBeuk, Cees H. – Journal of Educational Measurement, 1984
A systematic method for compromise between absolute and relative examination standards is proposed. The passing score is assumed to be related to expected pass rate through a simple linear function. Results define a function relating the percentage of successful candidates given a specified passing score to the passing score. (Author/DWH)
Descriptors: Achievement Tests, Cutting Scores, Foreign Countries, Mathematical Models
Peer reviewedWhitely, Susan E. – Journal of Educational Measurement, 1977
A debate concerning specific issues and the general usefulness of the Rasch latent trait test model is continued. Methods of estimation, necessary sample size, and the applicability of the model are discussed. (JKS)
Descriptors: Error of Measurement, Item Analysis, Mathematical Models, Measurement
Peer reviewedBudescu, David – Journal of Educational Measurement, 1985
An important determinant of equating process efficiency is the correlation between the anchor test and components of each form. Use of some monotonic function of this correlation as a measure of equating efficiency is suggested. A model relating anchor test length and test reliability to this measure of efficiency is presented. (Author/DWH)
Descriptors: Correlation, Equated Scores, Mathematical Models, Standardized Tests
Peer reviewedWoodruff, David – Journal of Educational Measurement, 1991
Improvements are made on previous estimates for the conditional standard error of measurement in prediction, the conditional standard error of estimation (CSEE), and the conditional standard error of prediction (CSEP). Better estimates of how test length affects CSEE and CSEP are derived. (SLD)
Descriptors: Equations (Mathematics), Error of Measurement, Estimation (Mathematics), Mathematical Models
Peer reviewedWright, Benjamin D. – Journal of Educational Measurement, 1977
Statements made in a previous article of this journal concerning the Rasch latent trait test model are questioned. Methods of estimation, necessary sample sizes, several formuli, and the general usefulness of the Rasch model are discussed. (JKS)
Descriptors: Computers, Error of Measurement, Item Analysis, Mathematical Models
Peer reviewedRoberts, Dennis M. – Journal of Educational Measurement, 1987
This study examines a score-difference model for the detection of cheating based on the difference between two scores for an examinee: one based on the appropriate scoring key and another based on an alternative, inappropriate key. It argues that the score-difference method could falsely accuse students as cheaters. (Author/JAZ)
Descriptors: Answer Keys, Cheating, Mathematical Models, Multiple Choice Tests
Peer reviewedVeale, James R.; Foreman, Dale I. – Journal of Educational Measurement, 1983
Statistical procedures for measuring heterogeneity of test item distractor distributions, or cultural variation, are presented. These procedures are based on the notion that examinees' responses to the incorrect options of a multiple-choice test provide more information concerning cultural bias than their correct responses. (Author/PN)
Descriptors: Ethnic Bias, Item Analysis, Mathematical Models, Multiple Choice Tests
Peer reviewedBaker, Frank B.; Al-Karni, Ali – Journal of Educational Measurement, 1991
Two methods of computing test equating coefficients under item response theory by the following authors are compared: (1) B. H. Loyd and H. D. Hoover (1980); and (2) M. L. Stocking and F. M. Lord (1983). Conditions under which the method of Stocking and Lord is preferable are described. (SLD)
Descriptors: Ability, College Entrance Examinations, Comparative Analysis, Equated Scores
Peer reviewedWainer, Howard; Lewis, Charles – Journal of Educational Measurement, 1990
Three different applications of the testlet concept are presented, and the psychometric models most suitable for each application are described. Difficulties that testlets can help overcome include (1) context effects; (2) item ordering; and (3) content balancing. Implications for test construction are discussed. (SLD)
Descriptors: Algorithms, Computer Assisted Testing, Elementary Secondary Education, Item Response Theory
Peer reviewedWainer, Howard – Journal of Educational Measurement, 1986
Describes recent research attempts to draw inferences about the relative standing of the states on the basis of mean SAT scores. This paper identifies five serious errors that call into question the validity of such inferences. Some plausible ways to avoid the errors are described. (Author/LMO)
Descriptors: College Entrance Examinations, Equated Scores, Mathematical Models, Predictor Variables
Peer reviewedSecolsky, Charles – Journal of Educational Measurement, 1983
A model is presented using examinee judgements in detecting ambiguous/misinterpreted items on teacher-made criterion-referenced tests. A computational example and guidelines for constructing domain categories and interpreting the indices are presented. (Author/PN)
Descriptors: Criterion Referenced Tests, Higher Education, Item Analysis, Mathematical Models
Peer reviewedKoffler, Stephen L. – Journal of Educational Measurement, 1980
Cut-off scores from two approaches for setting standards are examined. Standards determined from judgments about groups and from inspection of test content are compared. Results indicate that there was neither consistency nor pattern to cut-off scores set from the two procedures. (Author/RD)
Descriptors: Academic Standards, Cutting Scores, Educational Testing, Elementary Secondary Education


