NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 6,526 to 6,540 of 9,533 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Bridgeman, Brent; Cline, Frederick – Journal of Educational Measurement, 2004
Time limits on some computer-adaptive tests (CATs) are such that many examinees have difficulty finishing, and some examinees may be administered tests with more time-consuming items than others. Results from over 100,000 examinees suggested that about half of the examinees must guess on the final six questions of the analytical section of the…
Descriptors: Guessing (Tests), Timed Tests, Adaptive Testing, Computer Assisted Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Jee-Seon – Journal of Educational Measurement, 2006
Simulation and real data studies are used to investigate the value of modeling multiple-choice distractors on item response theory linking. Using the characteristic curve linking procedure for Bock's (1972) nominal response model presented by Kim and Hanson (2002), all-category linking (i.e., a linking based on all category characteristic curves…
Descriptors: Multiple Choice Tests, Test Items, Item Response Theory, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
DiStefano, Christine; Motl, Robert W. – Structural Equation Modeling: A Multidisciplinary Journal, 2006
This article used multitrait-multimethod methodology and covariance modeling for an investigation of the presence and correlates of method effects associated with negatively worded items on the Rosenberg Self-Esteem (RSE) scale (Rosenberg, 1989) using a sample of 757 adults. Results showed that method effects associated with negative item phrasing…
Descriptors: Adults, Correlation, Self Esteem, Surveys
Peer reviewed Peer reviewed
Direct linkDirect link
Beretvas, S. Natasha – Applied Psychological Measurement, 2004
In the bookmark standard-setting procedure, judges place "bookmarks" in a reordered test booklet containing items presented in order of increasing difficulty. Traditionally, the bookmark difficulty location (BDL) is on the trait continuum where, for dichotomous items, there is a two-thirds probability of a correct response and, for a score of "k"…
Descriptors: Probability, Standard Setting, Item Response Theory, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Hidalgo, M. Dolores; Lopez-Pina, Jose Antonio – Educational and Psychological Measurement, 2004
This article compares several procedures in their efficacy for detecting differential item functioning (DIF): logistic regression analysis, the Mantel-Haenszel (MH) procedure, and the modified Mantel-Haenszel procedure by Mazor, Clauser, and Hambleton. It also compares the effect size measures that these procedures provide. In this study,…
Descriptors: Effect Size, Test Bias, Regression (Statistics), Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Osterlind, Steven J.; Miao, Danmin; Sheng, Yanyan; Chia, Rosina C. – International Journal of Testing, 2004
This study investigated the interaction between different cultural groups and item type, and the ensuing effect on construct validity for a psychological inventory, the Myers-Briggs Type Indicator (MBTI, Form G). The authors analyzed 94 items from 2 Chinese-translated versions of the MBTI (Form G) for factorial differences among groups of…
Descriptors: Test Format, Undergraduate Students, Cultural Differences, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Levy, Roy; Mislevy, Robert J. – International Journal of Testing, 2004
The challenges of modeling students' performance in computer-based interactive assessments include accounting for multiple aspects of knowledge and skill that arise in different situations and the conditional dependencies among multiple aspects of performance. This article describes a Bayesian approach to modeling and estimating cognitive models…
Descriptors: Computer Assisted Testing, Markov Processes, Computer Networks, Bayesian Statistics
Peer reviewed Peer reviewed
Feldt, Leonard S. – Measurement & Evaluation in Counseling & Development, 2004
In some settings, the validity of a battery composite or a test score is enhanced by weighting some parts or items more heavily than others in the total score. This article describes methods of estimating the total score reliability coefficient when differential weights are used with items or parts.
Descriptors: Test Items, Scoring, Cognitive Processes, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Revuelta, Javier – Psychometrika, 2004
Two psychometric models are presented for evaluating the difficulty of the distractors in multiple-choice items. They are based on the criterion of rising distractor selection ratios, which facilitates interpretation of the subject and item parameters. Statistical inferential tools are developed in a Bayesian framework: modal a posteriori…
Descriptors: Multiple Choice Tests, Psychometrics, Models, Difficulty Level
Peer reviewed Peer reviewed
Direct linkDirect link
Alhija, Fadia Nasser-Abu; Wisenbaker, Joseph – Structural Equation Modeling: A Multidisciplinary Journal, 2006
A simulation study was conducted to examine the effect of item parceling on confirmatory factor analysis parameter estimates and their standard errors at different levels of sample size, number of indicators per factor, size of factor structure/pattern coefficients, magnitude of interfactor correlations, and variations in item-level data…
Descriptors: Monte Carlo Methods, Computation, Factor Analysis, Sample Size
Peer reviewed Peer reviewed
Direct linkDirect link
Penfield, Randall D. – Applied Measurement in Education, 2006
This study applied the maximum expected information (MEI) and the maximum posterior-weighted information (MPI) approaches of computer adaptive testing item selection to the case of a test using polytomous items following the partial credit model. The MEI and MPI approaches are described. A simulation study compared the efficiency of ability…
Descriptors: Bayesian Statistics, Adaptive Testing, Computer Assisted Testing, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Ariel, Adelaide; Veldkamp, Bernard P.; Breithaupt, Krista – Applied Psychological Measurement, 2006
Computerized multistage testing (MST) designs require sets of test questions (testlets) to be assembled to meet strict, often competing criteria. Rules that govern testlet assembly may dictate the number of questions on a particular subject or may describe desirable statistical properties for the test, such as measurement precision. In an MST…
Descriptors: Item Response Theory, Item Banks, Psychometrics, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Burns, Daniel J.; Martens, Nicholas J.; Bertoni, Alicia A.; Sweeney, Emily J.; Lividini, Michelle D. – Journal of Experimental Psychology: Learning, Memory, and Cognition, 2006
In a repeated testing paradigm, list items receiving item-specific processing are more likely to be recovered across successive tests (item gains), whereas items receiving relational processing are likely to be forgotten progressively less on successive tests. Moreover, analysis of cumulative-recall curves has shown that item-specific processing…
Descriptors: Item Analysis, Recall (Psychology), Cognitive Psychology, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Revuelta, Javier – Journal of Educational and Behavioral Statistics, 2004
This article presents a psychometric model for estimating ability and item-selection strategies in self-adapted testing. In contrast to computer adaptive testing, in self-adapted testing the examinees are allowed to select the difficulty of the items. The item-selection strategy is defined as the distribution of difficulty conditional on the…
Descriptors: Psychometrics, Adaptive Testing, Test Items, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Boo, Hong Kwen – Asia-Pacific Forum on Science Learning and Teaching, 2006
Assessment is an integral and vital part of teaching and learning, providing feedback on progress through the assessment period to both learners and teachers. However, if test items are flawed because of misconceptions held by the questions setter, then such test items are invalid as assessment tools. Moreover, such flawed items are also likely to…
Descriptors: Test Items, Science Tests, Water, Test Validity
Pages: 1  |  ...  |  432  |  433  |  434  |  435  |  436  |  437  |  438  |  439  |  440  |  ...  |  636