Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 0 |
| Since 2007 (last 20 years) | 8 |
Descriptor
| Simulation | 13 |
| Test Content | 13 |
| Computer Assisted Testing | 7 |
| Test Items | 7 |
| Comparative Analysis | 5 |
| Adaptive Testing | 4 |
| Item Response Theory | 4 |
| Test Construction | 4 |
| Scores | 3 |
| Test Format | 3 |
| Test Length | 3 |
| More ▼ | |
Source
Author
| Cui, Zhongmin | 2 |
| Kolen, Michael J. | 2 |
| Steffen, Manfred | 2 |
| Adams, Raymond J. | 1 |
| Babcock, Ben | 1 |
| Barrett, Richard S. | 1 |
| Berezner, Alla | 1 |
| Chang, Hua-Hua | 1 |
| Chiang, Claire | 1 |
| Chien, Yuehmei | 1 |
| Conrad, Kendon J. | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 11 |
| Reports - Research | 7 |
| Reports - Evaluative | 5 |
| Speeches/Meeting Papers | 2 |
| Opinion Papers | 1 |
| Tests/Questionnaires | 1 |
Education Level
| Elementary Secondary Education | 1 |
| Secondary Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
| Iowa Tests of Basic Skills | 1 |
| Program for International… | 1 |
What Works Clearinghouse Rating
Wyse, Adam E.; Babcock, Ben – Journal of Educational Measurement, 2016
A common suggestion made in the psychometric literature for fixed-length classification tests is that one should design tests so that they have maximum information at the cut score. Designing tests in this way is believed to maximize the classification accuracy and consistency of the assessment. This article uses simulated examples to illustrate…
Descriptors: Cutting Scores, Psychometrics, Test Construction, Classification
Adams, Raymond J.; Lietz, Petra; Berezner, Alla – Large-scale Assessments in Education, 2013
Background: While rotated test booklets have been employed in large-scale assessments to increase the content coverage of the assessments, rotation has not yet been applied to the context questionnaires administered to respondents. Methods: This paper describes the development of a methodology that uses rotated context questionnaires in…
Descriptors: Questionnaires, Item Response Theory, Foreign Countries, Achievement Tests
Shin, Chingwei David; Chien, Yuehmei; Way, Walter Denny – Pearson, 2012
Content balancing is one of the most important components in the computerized adaptive testing (CAT) especially in the K to 12 large scale tests that complex constraint structure is required to cover a broad spectrum of content. The purpose of this study is to compare the weighted penalty model (WPM) and the weighted deviation method (WDM) under…
Descriptors: Computer Assisted Testing, Elementary Secondary Education, Test Content, Models
Cui, Zhongmin; Kolen, Michael J. – Journal of Educational Measurement, 2009
This article considers two new smoothing methods in equipercentile equating, the cubic B-spline presmoothing method and the direct presmoothing method. Using a simulation study, these two methods are compared with established methods, the beta-4 method, the polynomial loglinear method, and the cubic spline postsmoothing method, under three sample…
Descriptors: Equated Scores, Methods, Sample Size, Test Content
Riley, Barth B.; Dennis, Michael L.; Conrad, Kendon J. – Applied Psychological Measurement, 2010
This simulation study sought to compare four different computerized adaptive testing (CAT) content-balancing procedures designed for use in a multidimensional assessment with respect to measurement precision, symptom severity classification, validity of clinical diagnostic recommendations, and sensitivity to atypical responding. The four…
Descriptors: Simulation, Computer Assisted Testing, Adaptive Testing, Comparative Analysis
Meyers, Jason L.; Miller, G. Edward; Way, Walter D. – Applied Measurement in Education, 2009
In operational testing programs using item response theory (IRT), item parameter invariance is threatened when an item appears in a different location on the live test than it did when it was field tested. This study utilizes data from a large state's assessments to model change in Rasch item difficulty (RID) as a function of item position change,…
Descriptors: Test Items, Test Content, Testing Programs, Simulation
Cui, Zhongmin; Kolen, Michael J. – Applied Psychological Measurement, 2008
This article considers two methods of estimating standard errors of equipercentile equating: the parametric bootstrap method and the nonparametric bootstrap method. Using a simulation study, these two methods are compared under three sample sizes (300, 1,000, and 3,000), for two test content areas (the Iowa Tests of Basic Skills Maps and Diagrams…
Descriptors: Test Length, Test Content, Simulation, Computation
Peer reviewedDavis, Laurie Laughlin; Pastor, Dena A.; Dodd, Barbara G.; Chiang, Claire; Fitzpatrick, Steven J. – Journal of Applied Measurement, 2003
Examined the effectiveness of the Sympson-Hetter technique and rotated content balancing relative to no exposure control and no content rotation conditions in a computerized adaptive testing system based on the partial credit model. Simulation results show the Sympson-Hetter technique can be used with minimal impact on measurement precision,…
Descriptors: Adaptive Testing, Computer Assisted Testing, Selection, Simulation
Peer reviewedLeung, Chi-Keung; Chang, Hua-Hua; Hau, Kit-Tai – Educational and Psychological Measurement, 2003
Studied three stratification designs for computerized adaptive testing in conjunction with three well-developed content balancing methods. Simulation study results show substantial differences in item overlap rate and pool utilization among different methods. Recommends an optimal combination of stratification design and content balancing method.…
Descriptors: Adaptive Testing, Computer Assisted Testing, Item Banks, Simulation
Rotou, Ourania; Patsula, Liane; Steffen, Manfred; Rizavi, Saba – ETS Research Report Series, 2007
Traditionally, the fixed-length linear paper-and-pencil (P&P) mode of administration has been the standard method of test delivery. With the advancement of technology, however, the popularity of administering tests using adaptive methods like computerized adaptive testing (CAT) and multistage testing (MST) has grown in the field of measurement…
Descriptors: Comparative Analysis, Test Format, Computer Assisted Testing, Models
Peer reviewedBarrett, Richard S. – Public Personnel Management, 1992
The Content Validation Form is presented as a means of proving that occupational tests provide a representative work sample or knowledge, skill, or ability necessary for a job. It is best used during test construction by a panel of subject matter experts. (SK)
Descriptors: Content Validity, Item Analysis, Multiple Choice Tests, Occupational Tests
Thomasson, Gary L. – 1997
Score comparability is important to those who take tests and those who use them. One important concept related to test score comparability is that of "equity," which is defined as existing when examinees are indifferent as to which of two alternate forms of a test they would prefer to take. By their nature, computerized adaptive tests…
Descriptors: Ability, Adaptive Testing, Comparative Analysis, Computer Assisted Testing
Robin, Frédéric; van der Linden, Wim J.; Eignor, Daniel R.; Steffen, Manfred; Stocking, Martha L. – ETS Research Report Series, 2005
The relatively new shadow test approach (STA) to computerized adaptive testing (CAT) proposed by Wim van der Linden is a potentially attractive alternative to the weighted deviation algorithm (WDA) implemented at ETS. However, it has not been evaluated under testing conditions representative of current ETS testing programs. Of interest was whether…
Descriptors: Test Construction, Computer Assisted Testing, Simulation, Evaluation Methods

Direct link
