Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 1 |
| Since 2007 (last 20 years) | 1 |
Descriptor
| Mathematical Models | 18 |
| Multiple Choice Tests | 18 |
| Test Reliability | 18 |
| Guessing (Tests) | 7 |
| Test Construction | 6 |
| Item Analysis | 5 |
| Higher Education | 4 |
| Scores | 4 |
| Scoring | 4 |
| Test Items | 4 |
| Achievement Tests | 3 |
| More ▼ | |
Source
| Educational and Psychological… | 2 |
| Applied Measurement in… | 1 |
| Applied Psychological… | 1 |
| Assessment & Evaluation in… | 1 |
| European Journal of… | 1 |
| Journal of Experimental… | 1 |
| Psychometrika | 1 |
Author
| Wilcox, Rand R. | 3 |
| Reckase, Mark D. | 2 |
| Burton, Richard F. | 1 |
| Feldt, Leonard S. | 1 |
| Gorbunova, Tatiana N. | 1 |
| Hamm, Debra W. | 1 |
| Hendrickson, Gerry F. | 1 |
| Izard, J. F. | 1 |
| Johnson, Eugene G. | 1 |
| Kane, Michael | 1 |
| Kane, Michael T. | 1 |
| More ▼ | |
Publication Type
| Reports - Research | 10 |
| Journal Articles | 7 |
| Reports - Evaluative | 3 |
| Speeches/Meeting Papers | 3 |
| Reports - Descriptive | 2 |
| Collected Works - General | 1 |
| Guides - General | 1 |
Education Level
| Higher Education | 1 |
| Postsecondary Education | 1 |
Audience
Location
| Russia | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| National Assessment of… | 1 |
What Works Clearinghouse Rating
Gorbunova, Tatiana N. – European Journal of Contemporary Education, 2017
The subject of the research is to build methodologies to evaluate the student knowledge by testing. The author points to the importance of feedback about the mastering level in the learning process. Testing is considered as a tool. The object of the study is to create the test system models for defence practice problems. Special attention is paid…
Descriptors: Testing, Evaluation Methods, Feedback (Response), Simulation
Peer reviewedRaju, Nambury S. – Educational and Psychological Measurement, 1982
A necessary and sufficient condition for a perfectly homogeneous test in the sense of Loevinger is stated and proved. Using this result, a formula for computing the maximum possible KR-20 when the test variance is assumed fixed is presented. A new index of test homogeneity is also presented and discussed. (Author/BW)
Descriptors: Mathematical Formulas, Mathematical Models, Multiple Choice Tests, Test Reliability
Peer reviewedWilcox, Rand R. – Educational and Psychological Measurement, 1982
Results in the engineering literature on "k out of n system reliability" can be used to characterize tests based on estimates of the probability of correctly determining whether the examinee knows the correct response. In particular, the minimum number of distractors required for multiple-choice tests can be empirically determined.…
Descriptors: Achievement Tests, Mathematical Models, Multiple Choice Tests, Test Format
Peer reviewedBurton, Richard F. – Assessment & Evaluation in Higher Education, 2001
Describes four measures of test unreliability that quantify effects of question selection and guessing, both separately and together--three chosen for immediacy and one for greater mathematical elegance. Quantifies their dependence on test length and number of answer options per question. Concludes that many multiple choice tests are unreliable…
Descriptors: Guessing (Tests), Mathematical Models, Multiple Choice Tests, Objective Tests
Koch, Bill R.; Reckase, Mark D. – 1978
A live tailored testing study was conducted to compare the results of using either the one-parameter logistic model or the three-parameter logistic model to measure the performance of college students on multiple choice vocabulary items. The results of the study showed the three-parameter tailored testing procedure to be superior to the…
Descriptors: Adaptive Testing, Comparative Analysis, Goodness of Fit, Higher Education
Peer reviewedZimmerman, Donald W.; And Others – Journal of Experimental Education, 1984
Three types of test were compared: a completion test, a matching test, and a multiple-choice test. The completion test was more reliable than the matching test, and the matching test was more reliable than the multiple-choice test. (Author/BW)
Descriptors: Comparative Analysis, Error of Measurement, Higher Education, Mathematical Models
Peer reviewedKane, Michael; Moloney, James – Applied Psychological Measurement, 1978
The answer-until-correct (AUC) procedure requires that examinees respond to a multi-choice item until they answer it correctly. Using a modified version of Horst's model for examinee behavior, this paper compares the effect of guessing on item reliability for the AUC procedure and the zero-one scoring procedure. (Author/CTM)
Descriptors: Guessing (Tests), Item Analysis, Mathematical Models, Multiple Choice Tests
Scheetz, James P.; vonFraunhofer, J. Anthony – 1980
Subkoviak suggested a technique for estimating both group reliability and the reliability associated with assigning a given individual to a mastery or non-mastery category based on a single test administration. Two assumptions underlie this model. First, it is assumed that had successive test administrations occurred, scores for each individual…
Descriptors: Criterion Referenced Tests, Cutting Scores, Error of Measurement, Higher Education
Peer reviewedFeldt, Leonard S. – Applied Measurement in Education, 1993
The recommendation that the reliability of multiple-choice tests will be enhanced if the distribution of item difficulties is concentrated at approximately 0.50 is reinforced and extended in this article by viewing the 0/1 item scoring as a dichotomization of an underlying normally distributed ability score. (SLD)
Descriptors: Ability, Difficulty Level, Guessing (Tests), Mathematical Models
Wilcox, Rand R. – 1981
These studies in test adequacy focus on two problems: procedures for estimating reliability, and techniques for identifying ineffective distractors. Fourteen papers are presented on recent advances in measuring achievement (a response to Molenaar); "an extension of the Dirichlet-multinomial model that allows true score and guessing to be…
Descriptors: Achievement Tests, Criterion Referenced Tests, Guessing (Tests), Mathematical Models
Ryan, Joseph P.; Hamm, Debra W. – 1976
A procedure is described for increasing the reliability of tests after they have been given and for developing shorter but more reliable tests. Eight tests administered to 200 graduate students studying educational research are analyzed. The analysis considers the original tests, the items loading on the first factor of the test, and the items…
Descriptors: Career Development, Factor Analysis, Factor Structure, Item Analysis
Wilcox, Rand R. – 1982
This document contains three papers from the Methodology Project of the Center for the Study of Evaluation. Methods for characterizing test accuracy are reported in the first two papers. "Bounds on the K Out of N Reliability of a Test, and an Exact Test for Hierarchically Related Items" describes and illustrates how an extension of a…
Descriptors: Educational Testing, Evaluation Methods, Guessing (Tests), Latent Trait Theory
PDF pending restorationKane, Michael T.; Moloney, James M. – 1976
The Answer-Until-Correct (AUC) procedure has been proposed in order to increase the reliability of multiple-choice items. A model for examinees' behavior when they must respond to each item until they answer it correctly is presented. An expression for the reliability of AUC items, as a function of the characteristics of the item and the scoring…
Descriptors: Guessing (Tests), Item Analysis, Mathematical Models, Multiple Choice Tests
Peer reviewedKim, Jwa K.; Nicewander, W. Alan – Psychometrika, 1993
Bias, standard error, and reliability of five ability estimators were evaluated using Monte Carlo estimates of the unknown conditional means and variances of the estimators. Results indicate that estimates based on Bayesian modal, expected a posteriori, and weighted likelihood estimators were reasonably unbiased with relatively small standard…
Descriptors: Ability, Bayesian Statistics, Equations (Mathematics), Error of Measurement
Koch, William R.; Reckase, Mark D. – 1979
Tailored testing procedures for achievement testing were applied in a situation that failed to meet some of the specifications generally considered to be necessary for tailored testing. Discrepancies from the appropriate conditions included the use of small samples for calibrating items, and the use of an item pool that was not designed to be…
Descriptors: Achievement Tests, Adaptive Testing, Educational Testing, Higher Education
Previous Page | Next Page ยป
Pages: 1 | 2

