NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 421 to 435 of 636 results Save | Export
Peer reviewed Peer reviewed
Cliff, Norman; And Others – Applied Psychological Measurement, 1979
Monte Carlo research with TAILOR, a program using implied orders as a basis for tailored testing, is reported. TAILOR typically required about half the available items to estimate, for each simulated examinee, the responses on the remainder. (Author/CTM)
Descriptors: Adaptive Testing, Computer Programs, Item Sampling, Nonparametric Statistics
Peer reviewed Peer reviewed
Gustafsson, Jan-Eric – Educational and Psychological Measurement, 1980
The statistically correct conditional maximum likelihood (CML) estimation method has not been used because of numerical problems. A solution is presented which allows a rapid computation of the CML esitmates also for long tests. CML has decisive advantages in the construction of statistical tests of goodness of fit. (Author/CP)
Descriptors: Goodness of Fit, Item Analysis, Latent Trait Theory, Mathematical Formulas
Peer reviewed Peer reviewed
Bell, Richard; Lumsden, James – Applied Psychological Measurement, 1980
The effect of test length on predictive validity is examined empirically. For four tests, the curve of validity against test length had a very gentle slope for the longer tests and all tests could be reduced by more than 60 percent without appreciable decreases in validity. (Author/BW)
Descriptors: Foreign Countries, High School Seniors, High Schools, Mathematical Models
Peer reviewed Peer reviewed
Cudeck, Robert; And Others – Applied Psychological Measurement, 1979
TAILOR, a computer program which implements an approach to tailored testing, was examined by Monte Carlo methods. The evaluation showed the procedure to be highly reliable and capable of reducing the required number of tests items by about one half. (Author/JKS)
Descriptors: Adaptive Testing, Computer Programs, Feasibility Studies, Item Analysis
Peer reviewed Peer reviewed
Wilcox, Rand R. – Educational and Psychological Measurement, 1979
A problem of considerable importance in certain educational settings is determining how many items to include on a mastery test. Applying ranking and selection procedures, a solution is given which includes as a special case all existing single-stage, non-Bayesian solutions based on a strong true-score model. (Author/JKS)
Descriptors: Criterion Referenced Tests, Mastery Tests, Nonparametric Statistics, Probability
Peer reviewed Peer reviewed
Wilcox, Rand R. – Journal of Educational Statistics, 1979
Methods are described for obtaining upper and lower bounds to both false-positive and false-negative decisions with a mastery test. These methods make no assumptions about the form of the true score distribution. (CTM)
Descriptors: Bayesian Statistics, Cutting Scores, Mastery Tests, Mathematical Formulas
Peer reviewed Peer reviewed
Christiansen, Neil D.; And Others – Educational and Psychological Measurement, 1996
The usefulness of examining the structural validity of scores on multidimensional measures using nested hierarchical model comparisons was evaluated in 2 studies using the Social Problem Solving Inventory (SPSI) with samples of 464 and 216 undergraduates. Results support the conceptual model of the SPSI. (SLD)
Descriptors: Comparative Analysis, Construct Validity, Higher Education, Interpersonal Relationship
Peer reviewed Peer reviewed
Frisby, Craig L.; Traffanstedt, Bobby K. – Journal of College Reading and Learning, 2003
Investigates the relationship between total scores on the California Critical Thinking Skills Test (CCTST) and the time taken to complete it. Finds that slower test takers obtained significantly higher scores. Discusses implications of these findings for college instruction. (SG)
Descriptors: Critical Thinking, Educational Research, High Schools, Higher Education
Peer reviewed Peer reviewed
Stark, Stephen; Drasgow, Fritz – Applied Psychological Measurement, 2002
Describes item response and information functions for the Zinnes and Griggs paired comparison item response theory (IRT) model (1974) and presents procedures for estimating stimulus and person parameters. Monte Carlo simulations show that at least 400 ratings are required to obtain reasonably accurate estimates of the stimulus parameters and their…
Descriptors: Comparative Analysis, Computer Simulation, Error of Measurement, Item Response Theory
Peer reviewed Peer reviewed
Grover, Cathy A.; And Others – Teaching of Psychology, 1989
Examines test scores of students who selected a chapter-by-chapter testing program, and those who chose unit (four-chapter) testing. Findings show neither option associated with superior performance, unlike earlier studies. Suggests instructors allow their students to choose their preferred testing option. (NL)
Descriptors: Academic Achievement, Attitudes, Comparative Analysis, Educational Research
Peer reviewed Peer reviewed
Meijer, Rob R.; And Others – Applied Psychological Measurement, 1994
The power of the nonparametric person-fit statistic, U3, is investigated through simulations as a function of item characteristics, test characteristics, person characteristics, and the group to which examinees belong. Results suggest conditions under which relatively short tests can be used for person-fit analysis. (SLD)
Descriptors: Difficulty Level, Group Membership, Item Response Theory, Nonparametric Statistics
Peer reviewed Peer reviewed
Woodruff, David – Journal of Educational Measurement, 1991
Improvements are made on previous estimates for the conditional standard error of measurement in prediction, the conditional standard error of estimation (CSEE), and the conditional standard error of prediction (CSEP). Better estimates of how test length affects CSEE and CSEP are derived. (SLD)
Descriptors: Equations (Mathematics), Error of Measurement, Estimation (Mathematics), Mathematical Models
Peer reviewed Peer reviewed
Noonan, Brian W.; And Others – Applied Psychological Measurement, 1992
Studied the extent to which three appropriateness indexes, Z(sub 3), ECIZ4, and W, are well standardized in a Monte Carlo study. The ECIZ4 most closely approximated a normal distribution, and its skewness and kurtosis were more stable and less affected by test length and item response theory model than the others. (SLD)
Descriptors: Comparative Analysis, Item Response Theory, Mathematical Models, Maximum Likelihood Statistics
Clements, Andrea D.; Rothenberg, Lori – Research in the Schools, 1996
Undergraduate psychology examinations from 48 schools were analyzed to determine the proportion of items at each level of Bloom's Taxonomy, item format, and test length. Analyses indicated significant relationships between item complexity and test length even when taking format into account. Use of higher items may be related to shorter tests,…
Descriptors: Classification, Difficulty Level, Educational Objectives, Higher Education
Stocking, Martha L. – 1994
As adaptive testing moves toward operational implementation in large scale testing programs, where it is important that adaptive tests be as parallel as possible to existing linear tests, a number of practical issues arise. This paper concerns three such issues. First, optimum item pool size is difficult to determine in advance of pool…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Banks, Standards
Pages: 1  |  ...  |  25  |  26  |  27  |  28  |  29  |  30  |  31  |  32  |  33  |  ...  |  43