NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 7 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Clauser, Jerome C.; Hambleton, Ronald K.; Baldwin, Peter – Educational and Psychological Measurement, 2017
The Angoff standard setting method relies on content experts to review exam items and make judgments about the performance of the minimally proficient examinee. Unfortunately, at times content experts may have gaps in their understanding of specific exam content. These gaps are particularly likely to occur when the content domain is broad and/or…
Descriptors: Scores, Item Analysis, Classification, Decision Making
Peer reviewed Peer reviewed
Direct linkDirect link
Kam, Chester Chun Seng – Educational and Psychological Measurement, 2016
To measure the response style of acquiescence, researchers recommend the use of at least 15 items with heterogeneous content. Such an approach is consistent with its theoretical definition and is a substantial improvement over traditional methods. Nevertheless, measurement of acquiescence can be enhanced by two additional considerations: first, to…
Descriptors: Test Items, Response Style (Tests), Test Content, Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Li, Xueming; Sireci, Stephen G. – Educational and Psychological Measurement, 2013
Validity evidence based on test content is of essential importance in educational testing. One source for such evidence is an alignment study, which helps evaluate the congruence between tested objectives and those specified in the curriculum. However, the results of an alignment study do not always sufficiently capture the degree to which a test…
Descriptors: Content Validity, Multidimensional Scaling, Data Analysis, Educational Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Breithaupt, Krista; Hare, Donovan R. – Educational and Psychological Measurement, 2007
Many challenges exist for high-stakes testing programs offering continuous computerized administration. The automated assembly of test questions to exactly meet content and other requirements, provide uniformity, and control item exposure can be modeled and solved by mixed-integer programming (MIP) methods. A case study of the computerized…
Descriptors: Testing Programs, Psychometrics, Certification, Accounting
Peer reviewed Peer reviewed
Leung, Chi-Keung; Chang, Hua-Hua; Hau, Kit-Tai – Educational and Psychological Measurement, 2003
Studied three stratification designs for computerized adaptive testing in conjunction with three well-developed content balancing methods. Simulation study results show substantial differences in item overlap rate and pool utilization among different methods. Recommends an optimal combination of stratification design and content balancing method.…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Banks, Simulation
Peer reviewed Peer reviewed
Ludlow, Larry H.; Bell, Karen N. – Educational and Psychological Measurement, 1996
Fifty education majors in two sections responded to an Attitudes toward Mathematics and Its Teaching (ATMAT) scale. Results with two psychometric models, classical true-score theory and the one-parameter Rasch model, supported the ATMAT's reliability, content and construct validity, and invariance over three time points. (SLD)
Descriptors: College Students, Construct Validity, Education Majors, Elementary Education
Peer reviewed Peer reviewed
Abedi, Jamal; Baker, Eva L. – Educational and Psychological Measurement, 1995
Results from a performance assessment in which 68 high school students wrote essays support the use of latent variable modeling for estimating reliability, concurrent validity, and generalizability of a scoring rubric. The latent variable modeling approach overcomes the limitations of certain conventional statistical techniques in handling…
Descriptors: Criteria, Essays, Estimation (Mathematics), Generalizability Theory