Publication Date
| In 2026 | 0 |
| Since 2025 | 8 |
| Since 2022 (last 5 years) | 23 |
| Since 2017 (last 10 years) | 23 |
| Since 2007 (last 20 years) | 23 |
Descriptor
Source
Author
| April L. Zenisky | 2 |
| Javier Suárez-Álvarez | 2 |
| Maria Elena Oliveri | 2 |
| Stephen G. Sireci | 2 |
| Abdullah Al Fraidan | 1 |
| Achim Goerres | 1 |
| Amit Sevak | 1 |
| Bengt Gerdin | 1 |
| Bianca A. Simonsmeier | 1 |
| Chandralekha Singh | 1 |
| Chia-Wen Chen | 1 |
| More ▼ | |
Publication Type
Education Level
| Secondary Education | 4 |
| Junior High Schools | 3 |
| Middle Schools | 3 |
| Adult Education | 2 |
| High Schools | 2 |
| Higher Education | 2 |
| Postsecondary Education | 2 |
| Elementary Education | 1 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
| ACT Assessment | 1 |
| California Critical Thinking… | 1 |
| Cornell Critical Thinking Test | 1 |
| Watson Glaser Critical… | 1 |
| Youth Risk Behavior Survey | 1 |
What Works Clearinghouse Rating
Yi-Ling Wu; Yao-Hsuan Huang; Chia-Wen Chen; Po-Hsi Chen – Journal of Educational Measurement, 2025
Multistage testing (MST), a variant of computerized adaptive testing (CAT), differs from conventional CAT in that it is adapted at the module level rather than at the individual item level. Typically, all examinees begin the MST with a linear test form in the first stage, commonly known as the routing stage. In 2020, Han introduced an innovative…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Format, Measurement
Monica Casella; Pasquale Dolce; Michela Ponticorvo; Nicola Milano; Davide Marocco – Educational and Psychological Measurement, 2024
Short-form development is an important topic in psychometric research, which requires researchers to face methodological choices at different steps. The statistical techniques traditionally used for shortening tests, which belong to the so-called exploratory model, make assumptions not always verified in psychological data. This article proposes a…
Descriptors: Artificial Intelligence, Test Construction, Test Format, Psychometrics
Mustafa Ilhan; Nese Güler; Gülsen Tasdelen Teker; Ömer Ergenekon – International Journal of Assessment Tools in Education, 2024
This study aimed to examine the effects of reverse items created with different strategies on psychometric properties and respondents' scale scores. To this end, three versions of a 10-item scale in the research were developed: 10 positive items were integrated in the first form (Form-P) and five positive and five reverse items in the other two…
Descriptors: Test Items, Psychometrics, Scores, Measures (Individuals)
Jing Ma – ProQuest LLC, 2024
This study investigated the impact of scoring polytomous items later on measurement precision, classification accuracy, and test security in mixed-format adaptive testing. Utilizing the shadow test approach, a simulation study was conducted across various test designs, lengths, number and location of polytomous item. Results showed that while…
Descriptors: Scoring, Adaptive Testing, Test Items, Classification
Cornelia Eva Neuert – Sociological Methods & Research, 2024
The quality of data in surveys is affected by response burden and questionnaire length. With an increasing number of questions, respondents can become bored, tired, and annoyed and may take shortcuts to reduce the effort needed to complete the survey. In this article, direct evidence is presented on how the position of items within a web…
Descriptors: Online Surveys, Test Items, Test Format, Test Construction
Morgan McCracken; Jonathan D. Bostic; Timothy D. Folger – TechTrends: Linking Research and Practice to Improve Learning, 2024
Assessment is central to teaching and learning, and recently there has been a substantive shift from paper-and-pencil assessments towards technology delivered assessments such as computer-adaptive tests. Fairness is an important aspect of the assessment process, including design, administration, test-score interpretation, and data utility. The…
Descriptors: Middle School Students, Student Attitudes, Culture Fair Tests, Mathematics Tests
Christian Berggren; Bengt Gerdin; Solmaz Filiz Karabag – Journal of Academic Ethics, 2025
The exposure of scientific scandals and the increase of dubious research practices have generated a stream of studies on Questionable Research Practices (QRPs), such as failure to acknowledge co-authors, selective presentation of findings, or removal of data not supporting desired outcomes. In contrast to high-profile fraud cases, QRPs can be…
Descriptors: Test Construction, Test Bias, Test Format, Response Style (Tests)
Jeff Allen; Jay Thomas; Stacy Dreyer; Scott Johanningmeier; Dana Murano; Ty Cruce; Xin Li; Edgar Sanchez – ACT Education Corp., 2025
This report describes the process of developing and validating the enhanced ACT. The report describes the changes made to the test content and the processes by which these design decisions were implemented. The authors describe how they shared the overall scope of the enhancements, including the initial blueprints, with external expert panels,…
Descriptors: College Entrance Examinations, Testing, Change, Test Construction
Jan Karem Höhne; Achim Goerres – International Journal of Social Research Methodology, 2024
The measurement of political solidarities and related concepts is an important endeavor in numerous scientific disciplines, such as political and social science research. European surveys, such as the Eurobarometer, frequently measure these concepts for people's home country and Europe raising questions with respect to the order of precedence.…
Descriptors: Surveys, Attitude Measures, Political Attitudes, Foreign Countries
Fu-Yun Yu – Interactive Learning Environments, 2024
Currently, 50 + learning systems supporting student question-generation (SQG) activities have been developed. While generating questions of different types is supported in many of these systems, systems allowing students to generate questions around a scenario (i.e. student testlet-generation, STG) are not yet available. Noting the increasing…
Descriptors: Computer Assisted Testing, Test Format, Test Construction, Test Items
Peter A. Edelsbrunner; Bianca A. Simonsmeier; Michael Schneider – Educational Psychology Review, 2025
Knowledge is an important predictor and outcome of learning and development. Its measurement is challenged by the fact that knowledge can be integrated and homogeneous, or fragmented and heterogeneous, which can change through learning. These characteristics of knowledge are at odds with current standards for test development, demanding a high…
Descriptors: Meta Analysis, Predictor Variables, Learning Processes, Knowledge Level
Judy R. Wilkerson; W. Steve Lang; LaSonya Moore – Journal of Research in Education, 2025
The DAATS (Dispositions Assessments Aligned with Teacher Standards) battery is a series of five instruments of different item types that measure teachers' consistency with the critical dispositions embedded in the InTASC Standards. The purpose of this study was to continue a 20-year research project on the development and implementation of…
Descriptors: Educational Assessment, National Standards, Teacher Evaluation, Teacher Competencies
Pasquale Anselmi; Jürgen Heller; Luca Stefanutti; Egidio Robusto; Giulia Barillari – Education and Information Technologies, 2025
Competence-based test development (CbTD) is a novel method for constructing tests that are as informative as possible about the competence state (the set of skills an individual masters) underlying the item responses. If desired, the tests can also be minimal, meaning that no item can be eliminated without reducing their informativeness. To…
Descriptors: Competency Based Education, Test Construction, Test Length, Usability
Wim J. van der Linden; Luping Niu; Seung W. Choi – Journal of Educational and Behavioral Statistics, 2024
A test battery with two different levels of adaptation is presented: a within-subtest level for the selection of the items in the subtests and a between-subtest level to move from one subtest to the next. The battery runs on a two-level model consisting of a regular response model for each of the subtests extended with a second level for the joint…
Descriptors: Adaptive Testing, Test Construction, Test Format, Test Reliability
Kyung-Mi O. – Language Testing in Asia, 2024
This study examines the efficacy of artificial intelligence (AI) in creating parallel test items compared to human-made ones. Two test forms were developed: one consisting of 20 existing human-made items and another with 20 new items generated with ChatGPT assistance. Expert reviews confirmed the content parallelism of the two test forms.…
Descriptors: Comparative Analysis, Artificial Intelligence, Computer Software, Test Items
Previous Page | Next Page »
Pages: 1 | 2
Peer reviewed
Direct link
