Publication Date
| In 2026 | 0 |
| Since 2025 | 74 |
| Since 2022 (last 5 years) | 509 |
| Since 2017 (last 10 years) | 1084 |
| Since 2007 (last 20 years) | 2603 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Researchers | 169 |
| Practitioners | 49 |
| Teachers | 32 |
| Administrators | 8 |
| Policymakers | 8 |
| Counselors | 4 |
| Students | 4 |
| Media Staff | 1 |
Location
| Turkey | 173 |
| Australia | 81 |
| Canada | 79 |
| China | 72 |
| United States | 56 |
| Taiwan | 44 |
| Germany | 43 |
| Japan | 41 |
| United Kingdom | 39 |
| Iran | 37 |
| Indonesia | 35 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 1 |
| Meets WWC Standards with or without Reservations | 1 |
| Does not meet standards | 1 |
Peer reviewedTatsuoka, Kikumi, K.; Tatsuoka, Maurice M. – Journal of Educational Statistics, 1982
Two indices for measuring the degree of conformity or consistency of an individual examinee's response pattern on a set of items are developed. The use of the indices for spotting aberrant response patterns of examinees is detailed. (Author/JKS)
Descriptors: Error of Measurement, Error Patterns, Goodness of Fit, Item Analysis
Peer reviewedUnnithan, N. Prabha; Scheuble, Laurie K. – Teaching Sociology, 1983
A version of Tobin's framework for analyzing newspaper news items was used and evaluated in a college-level introductory sociology course. The results showed that students viewed the analysis assignment positively, increased their media consumption as a result of it, and found it a useful technique for applying sociological concepts. (Author/RM)
Descriptors: Educational Research, Higher Education, Introductory Courses, Item Analysis
Peer reviewedFurnham, Adrian; Henderson, Monika – Journal of Clinical Psychology, 1982
Content analyzed four personality inventories, Eysenck Personality Questionnaire, Minnesota Multiphasic Personality Inventory, California Personality Inventory, and Edwards Personality Preference Schedule to determine whether each item in the four inventories contained situational, temporal, or comparative information. Wide differences appeared…
Descriptors: Behavior Patterns, Comparative Analysis, Data Analysis, Item Analysis
Peer reviewedFlexser, Arthur J. – Psychological Review, 1981
Contingency analyses have been employed to assess the degree to which outcomes of successive tests of corresponding items deviate from stochastic independence. A method of adjusting contingency tables to remove the effects of subject and item inhomogeneities, is presented. The method represents a partial solution to the "Simpson's…
Descriptors: Correlation, Expectancy Tables, Goodness of Fit, Item Analysis
Peer reviewedTerwilliger, James S.; Lele, Kaustubh – Journal of Educational Measurement, 1979
Different indices for the internal consistency, reproducibility, or homogeneity of a test are based upon highly similar conceptual frameworks. Illustrations are presented to demonstrate how the maximum and minimum values of KR20 are influenced by test difficulty and the shape of the distribution of test scores. (Author/CTM)
Descriptors: Difficulty Level, Item Analysis, Mathematical Formulas, Statistical Analysis
Peer reviewedBrunk, H. D. – Psychometrika, 1981
Bayesian techniques are adapted to the estimation of stimulus-response curves. Illustrative examples deal with estimation of person characteristic curves and item characteristic curves in the context of mental testing, and with estimation of a stimulus-response curve using data from a psychophysical experiment. (Author/JKS)
Descriptors: Bayesian Statistics, Item Analysis, Latent Trait Theory, Least Squares Statistics
Peer reviewedAustin, Joe Dan – Psychometrika, 1981
On distractor-identification tests students mark as many distractors as possible on each test item. A grading scale is developed for this type testing. The score is optimal in that it yields an unbiased estimate of the student's score as if no guessing had occurred. (Author/JKS)
Descriptors: Guessing (Tests), Item Analysis, Measurement Techniques, Scoring Formulas
Peer reviewedSmith, Jeffrey K. – Educational and Psychological Measurement, 1980
Weber contends that the use of Rasch analysis, principal components analysis, and classical test analysis shows that an instrument designed to measure a "bilevel dimensionality" in probability achievement measures a single latent trait. That interpretation and the use of Rasch and classical analysis to establish unidimensionality are…
Descriptors: Academic Achievement, Bayesian Statistics, Cognitive Processes, Item Analysis
Peer reviewedBurkhart, Barry R.; And Others – Journal of Clinical Psychology, 1980
Obvious items were good predictors of all criteria; neutral items overlapped considerably with obvious items; and subtle items generally did not contribute uniquely to the prediction of any of the concurrent measures of depression. (Author)
Descriptors: Depression (Psychology), Item Analysis, Personality Measures, Predictive Validity
Peer reviewedPlake, Barbara S.; Hoover, H. D. – Journal of Experimental Education, 1979
A follow-up technique is needed to identify items contributing to items-by-groups interaction when using an ANOVA procedure to examine a test for biased items. The method described includes distribution theory for assessing level of significance and is sensitive to items at all difficulty levels. (Author/GSK)
Descriptors: Analysis of Variance, Goodness of Fit, Item Analysis, Statistical Bias
Peer reviewedHarris, Chester W.; Pearlman, Andrea Pastorok – Journal of Educational Statistics, 1978
A theory and a procedure are presented for estimating a domain parameter and item parameters for test items in a homogeneous domain, such that the combined domain and item parameters account for observed proportions right for each item in a test. (CTM)
Descriptors: Achievement Tests, Difficulty Level, Item Analysis, Mathematical Models
Peer reviewedQuereshi, M. Y.; Fisher, Thomas L. – Educational and Psychological Measurement, 1977
Logical estimates of item difficulty made by judges were compared to empirical estimates derived from a test administration. Results indicated substantial correspondence between logical and empirical estimates, and substantial variation among judges. Further, the more elaborate the system used by judges to make estimates, the more accurate the…
Descriptors: Court Judges, Difficulty Level, Evaluation Methods, Item Analysis
Peer reviewedRosenbaum, Paul R. – Psychometrika, 1988
Two theorems of unidimensional item response theory are extended to describe observable item response distributions when there is conditional independence between but not necessarily within item bundles. An item bundle is a small group of multiple-choice items sharing a common reading passage or a group of items sharing distractors. (SLD)
Descriptors: Equations (Mathematics), Item Analysis, Latent Trait Theory, Multiple Choice Tests
Peer reviewedJacobs, Stanley S. – Measurement and Evaluation in Counseling and Development, 1999
Comparability of two forms of the CCTST is evaluated. Data are compared with a previously reported study; relationships to comparable tests are considered. Evidence indicates low internal-consistency reliability estimates, a lack of comparability, and poor construct validity for Forms A and B of the CCTST. (Author/EMK)
Descriptors: College Students, Construct Validity, Critical Thinking, Higher Education
Peer reviewedKim, Mikyung – Language Testing, 2001
Investigates differential item functioning (DIF) across two different broad language groupings, Asian and European, in a speaking test in which the test takers' responses were rated polytomously. Data were collected from 1038 nonnative speakers of English from France, Hong Kong, Japan, Spain, Switzerland, and Thailand who took the SPEAK test in…
Descriptors: English (Second Language), Foreign Countries, Item Analysis, Language Tests


