Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 1 |
| Since 2007 (last 20 years) | 3 |
Descriptor
| Selection | 14 |
| Test Format | 14 |
| Test Items | 8 |
| Item Response Theory | 6 |
| Adaptive Testing | 5 |
| Computer Assisted Testing | 5 |
| Test Construction | 5 |
| Models | 4 |
| Comparative Analysis | 3 |
| High School Students | 3 |
| High Schools | 3 |
| More ▼ | |
Source
| Applied Psychological… | 1 |
| Assessment & Evaluation in… | 1 |
| Journal of Educational… | 1 |
| Measurement:… | 1 |
| Psychological Assessment | 1 |
| Research in the Schools | 1 |
Author
Publication Type
| Reports - Research | 8 |
| Journal Articles | 6 |
| Reports - Evaluative | 6 |
| Speeches/Meeting Papers | 4 |
Education Level
| Elementary Secondary Education | 1 |
| Grade 12 | 1 |
| Grade 4 | 1 |
| Grade 8 | 1 |
| Higher Education | 1 |
| Postsecondary Education | 1 |
Audience
Location
| Texas | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| Advanced Placement… | 1 |
| National Assessment of… | 1 |
| Texas Assessment of Academic… | 1 |
| Wechsler Adult Intelligence… | 1 |
What Works Clearinghouse Rating
Luo, Yong – Measurement: Interdisciplinary Research and Perspectives, 2021
To date, only frequentist model-selection methods have been studied with mixed-format data in the context of IRT model-selection, and it is unknown how popular Bayesian model-selection methods such as DIC, WAIC, and LOO perform. In this study, we present the results of a comprehensive simulation study that compared the performances of eight…
Descriptors: Item Response Theory, Test Format, Selection, Methods
Irwin, Brian; Hepplestone, Stuart – Assessment & Evaluation in Higher Education, 2012
There have been calls in the literature for changes to assessment practices in higher education, to increase flexibility and give learners more control over the assessment process. This article explores the possibilities of allowing student choice in the format used to present their work, as a starting point for changing assessment, based on…
Descriptors: Student Evaluation, College Students, Selection, Computer Assisted Testing
Whittaker, Tiffany A.; Chang, Wanchen; Dodd, Barbara G. – Applied Psychological Measurement, 2012
When tests consist of multiple-choice and constructed-response items, researchers are confronted with the question of which item response theory (IRT) model combination will appropriately represent the data collected from these mixed-format tests. This simulation study examined the performance of six model selection criteria, including the…
Descriptors: Item Response Theory, Models, Selection, Criteria
Berger, Martijn P. F.; Veerkamp, Wim J. J. – 1994
The designing of tests has been a source of concern for test developers over the past decade. Various kinds of test forms have been applied. Among these are the fixed-form test, the adaptive test, and the testlet. Each of these forms has its own design. In this paper, the construction of test forms is placed within the general framework of optimal…
Descriptors: Adaptive Testing, Foreign Countries, Research Design, Selection
Peer reviewedWard, L. Charles; Ryan, Joseph J. – Psychological Assessment, 1996
Validity and reliability were calculated from data in the standardization sample of the Wechsler Adult Intelligence Scale--Revised for 565 proposed short forms. Time saved in comparison with use of the long form was estimated. The most efficient combinations were generally those composed of subtests that were quick to administer. (SLD)
Descriptors: Cost Effectiveness, Intelligence Tests, Selection, Test Format
Stocking, Martha L.; Lewis, Charles – 1995
In the periodic testing environment associated with conventional paper-and-pencil tests, the frequency with which items are seen by test-takers is tightly controlled in advance of testing by policies that regulate both the reuse of test forms and the frequency with which candidates may take the test. In the continuous testing environment…
Descriptors: Adaptive Testing, Computer Assisted Testing, Selection, Test Construction
Wang, Yu-Chung Lawrence – 1994
The first purpose of this study was to investigate the stability of two essential dimensionality measures across 10 random samples within a particular assessment item (AT1) selection. Other purposes were to investigate the discrepancy of the essential unidimensionality estimates for a test across different AT1 selections and sample sizes and to…
Descriptors: Correlation, Educational Assessment, Estimation (Mathematics), Item Response Theory
Barnette, J. Jackson – Research in the Schools, 2001
Studied the primacy effect (tendency to select items closer to the left side of the response scale) in Likert scales worded from "Strongly Disagree" to "Strongly Agree" and in the opposite direction. Findings for 386 high school and college students show no primacy effect, although negatively worded stems had an effect on Cronbach's alpha. (SLD)
Descriptors: College Students, High School Students, High Schools, Higher Education
Peer reviewedWainer, Howard; And Others – Journal of Educational Measurement, 1994
The comparability of scores on test forms that are constructed through examinee item choice is examined in an item response theory framework. The approach is illustrated with data from the College Board's Advanced Placement Test in Chemistry taken by over 18,000 examinees. (SLD)
Descriptors: Advanced Placement, Chemistry, Comparative Analysis, Constructed Response
Bizot, Elizabeth B.; Goldman, Steven H. – 1994
A study was conducted to evaluate the effects of choice of item response theory (IRT) model, parameter calibration group, starting ability estimate, and stopping criterion on the conversion of an 80-item vocabulary test to computer adaptive format. Three parameter calibration groups were tested: (1) a group of 1,000 high school seniors, (2) a…
Descriptors: Ability, Adaptive Testing, Computer Assisted Testing, Estimation (Mathematics)
Stocking, Martha L. – 1988
The construction of parallel editions of conventional tests for purposes of test security while maintaining score comparability has always been a recognized and difficult problem in psychometrics and test construction. The introduction of new modes of test construction, e.g., adaptive testing, changes the nature of the problem, but does not make…
Descriptors: Adaptive Testing, Comparative Analysis, Computer Assisted Testing, Identification
Wainer, Howard; And Others – 1991
When an examination consists, in whole or in part, of constructed response items, it is a common practice to allow the examinee to choose among a variety of questions. This procedure is usually adopted so that the limited number of items that can be completed in the allotted time does not unfairly affect the examinee. This results in the de facto…
Descriptors: Adaptive Testing, Chemistry, Comparative Analysis, Computer Assisted Testing
Fan, Xitao; And Others – 1994
The hypothesis that faulty classical psychometric and sampling procedures in test construction could generate systematic bias against ethnic groups with smaller representation in the test construction sample was studied empirically. Two test construction models were developed: one with differential representation of ethnic groups (White, African…
Descriptors: Ethnic Groups, Genetics, High School Students, High Schools
Dirir, Mohamed A. – 1995
The effectiveness of an optimal item selection method in designing parallel test forms was studied during the development of two forms that were parallel to an existing form for each of three language arts tests for fourth graders used in the Connecticut Mastery Test. Two listening comprehension forms, two reading comprehension forms, and two…
Descriptors: Elementary School Students, Grade 4, Intermediate Grades, Item Banks

Direct link
