Publication Date
| In 2026 | 0 |
| Since 2025 | 1 |
| Since 2022 (last 5 years) | 4 |
| Since 2017 (last 10 years) | 9 |
| Since 2007 (last 20 years) | 13 |
Descriptor
| Pretesting | 13 |
| Test Items | 13 |
| Test Construction | 7 |
| Adaptive Testing | 4 |
| Computer Assisted Testing | 4 |
| Item Banks | 4 |
| Item Response Theory | 4 |
| Statistical Analysis | 4 |
| Comparative Analysis | 3 |
| Computation | 3 |
| Foreign Countries | 3 |
| More ▼ | |
Source
Author
| Adams, Betty A. J. | 1 |
| Benjamin W. Domingue | 1 |
| Choe, Edison M. | 1 |
| Cobern, William W. | 1 |
| Davey, Tim | 1 |
| Dorans, Neil J. | 1 |
| Ersen, Rabia Karatoprak | 1 |
| Guo, Hongwen | 1 |
| Hilton, Charlotte Emma | 1 |
| Howard, Matt C. | 1 |
| Jones, Emily | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 13 |
| Reports - Research | 11 |
| Tests/Questionnaires | 2 |
| Reports - Descriptive | 1 |
| Reports - Evaluative | 1 |
Education Level
| Higher Education | 4 |
| Postsecondary Education | 3 |
| Elementary Education | 2 |
| Early Childhood Education | 1 |
| Grade 2 | 1 |
| Intermediate Grades | 1 |
| Primary Education | 1 |
Audience
| Practitioners | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| Graduate Record Examinations | 1 |
| SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Ersen, Rabia Karatoprak; Lee, Won-Chan – Journal of Educational Measurement, 2023
The purpose of this study was to compare calibration and linking methods for placing pretest item parameter estimates on the item pool scale in a 1-3 computerized multistage adaptive testing design in terms of item parameter recovery. Two models were used: embedded-section, in which pretest items were administered within a separate module, and…
Descriptors: Pretesting, Test Items, Computer Assisted Testing, Adaptive Testing
Bayesian Logistic Regression: A New Method to Calibrate Pretest Items in Multistage Adaptive Testing
TsungHan Ho – Applied Measurement in Education, 2023
An operational multistage adaptive test (MST) requires the development of a large item bank and the effort to continuously replenish the item bank due to concerns about test security and validity over the long term. New items should be pretested and linked to the item bank before being used operationally. The linking item volume fluctuations in…
Descriptors: Bayesian Statistics, Regression (Statistics), Test Items, Pretesting
Lim, Hwanggyu; Choe, Edison M. – Journal of Educational Measurement, 2023
The residual differential item functioning (RDIF) detection framework was developed recently under a linear testing context. To explore the potential application of this framework to computerized adaptive testing (CAT), the present study investigated the utility of the RDIF[subscript R] statistic both as an index for detecting uniform DIF of…
Descriptors: Test Items, Computer Assisted Testing, Item Response Theory, Adaptive Testing
Cobern, William W.; Adams, Betty A. J. – International Journal of Assessment Tools in Education, 2020
What follows is a practical guide for establishing the validity of a survey for research purposes. The motivation for providing this guide is our observation that researchers, not necessarily being survey researchers per se, but wanting to use a survey method, lack a concise resource on validity. There is far more to know about surveys and survey…
Descriptors: Surveys, Test Validity, Test Construction, Test Items
Joshua B. Gilbert; Luke W. Miratrix; Mridul Joshi; Benjamin W. Domingue – Journal of Educational and Behavioral Statistics, 2025
Analyzing heterogeneous treatment effects (HTEs) plays a crucial role in understanding the impacts of educational interventions. A standard practice for HTE analysis is to examine interactions between treatment status and preintervention participant characteristics, such as pretest scores, to identify how different groups respond to treatment.…
Descriptors: Causal Models, Item Response Theory, Statistical Inference, Psychometrics
Howard, Matt C. – Practical Assessment, Research & Evaluation, 2018
Scale pretests analyze the suitability of individual scale items for further analysis, whether through judging their face validity, wording concerns, and/or other aspects. The current article reviews scale pretests, separated by qualitative and quantitative methods, in order to identify the differences, similarities, and even existence of the…
Descriptors: Pretesting, Measures (Individuals), Test Items, Statistical Analysis
Hilton, Charlotte Emma – International Journal of Social Research Methodology, 2017
The development of questionnaires, surveys and psychometric scales is an iterative research process that includes a number of carefully planned stages. Pretesting is a method of checking that questions work as intended and are understood by those individuals who are likely to respond to them. However, detailed reports of appropriate methods to…
Descriptors: Questionnaires, Pretesting, Interviews, Test Construction
Kim, Sooyeon; Robin, Frederic – ETS Research Report Series, 2017
In this study, we examined the potential impact of item misfit on the reported scores of an admission test from the subpopulation invariance perspective. The target population of the test consisted of 3 major subgroups with different geographic regions. We used the logistic regression function to estimate item parameters of the operational items…
Descriptors: Scores, Test Items, Test Bias, International Assessment
Wheadon, Jacob; Wright, Geoff A.; West, Richard E.; Skaggs, Paul – Journal of Technology Education, 2017
This study discusses the need, development, and validation of the Innovation Test Instrument (ITI). This article outlines how the researchers identified the content domain of the assessment and created test items. Then, it describes initial validation testing of the instrument. The findings suggest that the ITI is a good first step in creating an…
Descriptors: Innovation, Program Validation, Evaluation Needs, Test Construction
Liu, Jinghua; Guo, Hongwen; Dorans, Neil J. – ETS Research Report Series, 2014
Maintaining score interchangeability and scale consistency is crucial for any testing programs that administer multiple forms across years. The use of a multiple linking design, which involves equating a new form to multiple old forms and averaging the conversions, has been proposed to control scale drift. However, the use of multiple linking…
Descriptors: Comparative Analysis, Reliability, Test Construction, Equated Scores
Davey, Tim; Lee, Yi-Hsuan – ETS Research Report Series, 2011
Both theoretical and practical considerations have led the revision of the Graduate Record Examinations® (GRE®) revised General Test, here called the rGRE, to adopt a multistage adaptive design that will be continuously or nearly continuously administered and that can provide immediate score reporting. These circumstances sharply constrain the…
Descriptors: Context Effect, Scoring, Equated Scores, College Entrance Examinations
Sapriati, Amalia; Zuhairi, Aminudin – Turkish Online Journal of Distance Education, 2010
This paper addresses the use of computer-based testing in distance education, based on the experience of Universitas Terbuka (UT), Indonesia. Computer-based testing has been developed at UT for reasons of meeting the specific needs of distance students as the following: (1) students' inability to sit for the scheduled test; (2) conflicting test…
Descriptors: Alternative Assessment, Distance Education, Computer Assisted Testing, Computer System Design
Pyle, Katie; Jones, Emily; Williams, Chris; Morrison, Jo – Educational Research, 2009
Background: All national curriculum tests in England are pre-tested as part of the development process. Differences in pupil performance between pre-test and live test are consistently found. This difference has been termed the pre-test effect. Understanding the pre-test effect is essential in the test development and selection processes and in…
Descriptors: Foreign Countries, Pretesting, Context Effect, National Curriculum

Peer reviewed
Direct link
