Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 1 |
| Since 2007 (last 20 years) | 7 |
Descriptor
| Error of Measurement | 26 |
| Scoring Formulas | 26 |
| Test Reliability | 9 |
| True Scores | 7 |
| Test Items | 6 |
| Cutting Scores | 5 |
| Scores | 5 |
| Statistical Analysis | 5 |
| Multiple Choice Tests | 4 |
| Scoring | 4 |
| Test Theory | 4 |
| More ▼ | |
Source
Author
| Huynh, Huynh | 2 |
| Livingston, Samuel A. | 2 |
| Attali, Yigal | 1 |
| Bakker, J. | 1 |
| Bardhoshi, Gerta | 1 |
| Beek, F. J. A. | 1 |
| Berger, Dale E. | 1 |
| Brennan, Robert L. | 1 |
| Burton, Richard F. | 1 |
| Cureton, Edward E. | 1 |
| Dimoliatis, Ioannis D. K. | 1 |
| More ▼ | |
Publication Type
| Reports - Research | 16 |
| Journal Articles | 13 |
| Reports - Evaluative | 4 |
| Speeches/Meeting Papers | 4 |
| Opinion Papers | 1 |
| Reports - Descriptive | 1 |
| Tests/Questionnaires | 1 |
Education Level
| Higher Education | 3 |
| Postsecondary Education | 3 |
| Elementary Secondary Education | 2 |
Audience
| Researchers | 2 |
Location
| Japan | 1 |
| Mississippi | 1 |
| Pennsylvania | 1 |
| Turkey | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| Comprehensive Tests of Basic… | 1 |
What Works Clearinghouse Rating
Bardhoshi, Gerta; Erford, Bradley T. – Measurement and Evaluation in Counseling and Development, 2017
Precision is a key facet of test development, with score reliability determined primarily according to the types of error one wants to approximate and demonstrate. This article identifies and discusses several primary forms of reliability estimation: internal consistency (i.e., split-half, KR-20, a), test-retest, alternate forms, interscorer, and…
Descriptors: Scores, Test Reliability, Accuracy, Pretests Posttests
Ravesloot, C. J.; Van der Schaaf, M. F.; Muijtjens, A. M. M.; Haaring, C.; Kruitwagen, C. L. J. J.; Beek, F. J. A.; Bakker, J.; Van Schaik, J.P.J.; Ten Cate, Th. J. – Advances in Health Sciences Education, 2015
Formula scoring (FS) is the use of a don't know option (DKO) with subtraction of points for wrong answers. Its effect on construct validity and reliability of progress test scores, is subject of discussion. Choosing a DKO may not only be affected by knowledge level, but also by risk taking tendency, and may thus introduce construct-irrelevant…
Descriptors: Scoring Formulas, Tests, Scores, Construct Validity
Temel, Gülhan Orekici; Erdogan, Semra; Selvi, Hüseyin; Kaya, Irem Ersöz – Educational Sciences: Theory and Practice, 2016
Studies based on longitudinal data focus on the change and development of the situation being investigated and allow for examining cases regarding education, individual development, cultural change, and socioeconomic improvement in time. However, as these studies require taking repeated measures in different time periods, they may include various…
Descriptors: Investigations, Sample Size, Longitudinal Studies, Interrater Reliability
Holster, Trevor A.; Lake, J. – Language Assessment Quarterly, 2016
Stewart questioned Beglar's use of Rasch analysis of the Vocabulary Size Test (VST) and advocated the use of 3-parameter logistic item response theory (3PLIRT) on the basis that it models a non-zero lower asymptote for items, often called a "guessing" parameter. In support of this theory, Stewart presented fit statistics derived from…
Descriptors: Guessing (Tests), Item Response Theory, Vocabulary, Language Tests
Dimoliatis, Ioannis D. K.; Jelastopulu, Eleni – Universal Journal of Educational Research, 2013
The surgical theatre educational environment measures STEEM, OREEM and mini-STEEM for students (student-STEEM) comprise an up to now disregarded systematic overestimation (OE) due to inaccurate percentage calculation. The aim of the present study was to investigate the magnitude of and suggest a correction for this systematic bias. After an…
Descriptors: Educational Environment, Scores, Grade Prediction, Academic Standards
Peer reviewedPandey, Tej N.; Shoemaker, David M. – Educational and Psychological Measurement, 1975
Described herein are formulas and computational procedures for estimating the mean and second through fourth central moments of universe scores through multiple matrix sampling. Additionally, procedures are given for approximating the standard error associated with each estimate. All procedures are applicable when items are scored either…
Descriptors: Error of Measurement, Item Sampling, Matrices, Scoring Formulas
Peer reviewedRowley, Glenn L.; Traub, Ross E. – Journal of Educational Measurement, 1977
The consequences of formula scoring versus number right scoring are examined in relation to the assumptions commonly made about the behavior of examinees in testing situations. The choice between the two is shown to be dependent upon having reduced error variance or unbiasedness as a goal. (Author/JKS)
Descriptors: Error of Measurement, Scoring Formulas, Statistical Bias, Test Wiseness
Peer reviewedCureton, Edward E. – Educational and Psychological Measurement, 1971
A rebuttal of Frary's 1969 article in Educational and Psychological Measurement. (MS)
Descriptors: Error of Measurement, Guessing (Tests), Multiple Choice Tests, Scoring Formulas
Livingston, Samuel A. – 1981
The standard error of measurement (SEM) is a measure of the inconsistency in the scores of a particular group of test-takers. It is largest for test-takers with scores ranging in the 50 percent correct bracket; with nearly perfect scores, it is smaller. On tests used to make pass/fail decisions, the test-takers' scores tend to cluster in the range…
Descriptors: Error of Measurement, Estimation (Mathematics), Mathematical Formulas, Pass Fail Grading
Doppelt, Jerome E. – Test Service Bulletin, 1956
The standard error of measurement as a means for estimating the margin of error that should be allowed for in test scores is discussed. The true score measures the performance that is characteristic of the person tested; the variations, plus and minus, around the true score describe a characteristic of the test. When the standard deviation is used…
Descriptors: Bulletins, Error of Measurement, Measurement Techniques, Reliability
Peer reviewedMitchelmore, M. C. – British Journal of Educational Psychology, 1981
This paper presents a scientific rationale for deciding the number of points to use on a grading scale in any given assessment situation. The rationale is applied to two common methods of assessment (multiple-choice and essay tests) and an example of a composite assessment. (Author/SJL)
Descriptors: Error of Measurement, Essay Tests, Grading, Higher Education
Peer reviewedHarris, Chester W. – Journal of Educational Measurement, 1973
A brief note presenting algebraically equivalent formulas for the variances of three error types. (Author)
Descriptors: Algebra, Analysis of Covariance, Analysis of Variance, Error of Measurement
Peer reviewedMcGaw, Barry; Glass, Gene V. – American Educational Research Journal, 1980
There are difficulties in expressing effect sizes on a common metric when some studies use transformed scales to express group differences, or use factorial designs or covariance adjustments to obtain a reduced error term. A common metric on which effect sizes may be standardized is described. (Author/RL)
Descriptors: Control Groups, Error of Measurement, Mathematical Models, Research Problems
Peer reviewedKleven, Thor Arnfinn – Scandinavian Journal of Educational Research, 1979
Supposing different values of the standard measurement error, the relation of scale coarseness to the total amount of error is studied on the basis of probability distribution of error. The analyses are performed within two models of error and with two criteria of amount of error. (Editor/SJL)
Descriptors: Cutting Scores, Error of Measurement, Goodness of Fit, Grading
Attali, Yigal – ETS Research Report Series, 2007
Because there is no commonly accepted view of what makes for good writing, automated essay scoring (AES) ideally should be able to accommodate different theoretical positions, certainly at the level of state standards but also perhaps among teachers at the classroom level. This paper presents a practical approach and an interactive computer…
Descriptors: Computer Assisted Testing, Automation, Essay Tests, Scoring
Previous Page | Next Page »
Pages: 1 | 2
Direct link
