NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1 to 15 of 3,123 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Tom Benton – Practical Assessment, Research & Evaluation, 2025
This paper proposes an extension of linear equating that may be useful in one of two fairly common assessment scenarios. One is where different students have taken different combinations of test forms. This might occur, for example, where students have some free choice over the exam papers they take within a particular qualification. In this…
Descriptors: Equated Scores, Test Format, Test Items, Computation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Bin Tan; Nour Armoush; Elisabetta Mazzullo; Okan Bulut; Mark J. Gierl – International Journal of Assessment Tools in Education, 2025
This study reviews existing research on the use of large language models (LLMs) for automatic item generation (AIG). We performed a comprehensive literature search across seven research databases, selected studies based on predefined criteria, and summarized 60 relevant studies that employed LLMs in the AIG process. We identified the most commonly…
Descriptors: Artificial Intelligence, Test Items, Automation, Test Format
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Meltem Acar Güvendir; Seda Donat Bacioglu; Hasan Özgür; Sefa Uyanik; Fatmagül Gürbüz Akçay; Emre Güvendir – International Journal of Psychology and Educational Studies, 2025
Different types of test items influence students' test anxiety, and physiological measures such as heart rate provide a means of measuring this anxiety. This study aimed to explore the connection between test anxiety and examination item formats. It centered on 20 junior university students in Western Türkiye. The research monitored students'…
Descriptors: Foreign Countries, Test Anxiety, Measurement Techniques, Physiology
Peer reviewed Peer reviewed
Direct linkDirect link
Hao Lei; Libing Chen; Ming Ming Chiu; Longyue Fang; Yuxin Ding – Educational Psychology Review, 2025
Adding illustrations to texts might improve students' science achievement. This meta-analysis of 121 effect sizes from 63 studies of 7,621 students across five decades determines both the overall effect and moderators that account for differences across studies. Our random-effects model shows a positive effect of adding illustrations to texts on…
Descriptors: Illustrations, Textbooks, Science Achievement, Effect Size
Peer reviewed Peer reviewed
Direct linkDirect link
Yi-Ling Wu; Yao-Hsuan Huang; Chia-Wen Chen; Po-Hsi Chen – Journal of Educational Measurement, 2025
Multistage testing (MST), a variant of computerized adaptive testing (CAT), differs from conventional CAT in that it is adapted at the module level rather than at the individual item level. Typically, all examinees begin the MST with a linear test form in the first stage, commonly known as the routing stage. In 2020, Han introduced an innovative…
Descriptors: Computer Assisted Testing, Adaptive Testing, Test Format, Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Sohee Kim; Ki Lynn Cole – International Journal of Testing, 2025
This study conducted a comprehensive comparison of Item Response Theory (IRT) linking methods applied to a bifactor model, examining their performance on both multiple choice (MC) and mixed format tests within the common item nonequivalent group design framework. Four distinct multidimensional IRT linking approaches were explored, consisting of…
Descriptors: Item Response Theory, Comparative Analysis, Models, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Stefanie A. Wind; Yuan Ge – Measurement: Interdisciplinary Research and Perspectives, 2024
Mixed-format assessments made up of multiple-choice (MC) items and constructed response (CR) items that are scored using rater judgments include unique psychometric considerations. When these item types are combined to estimate examinee achievement, information about the psychometric quality of each component can depend on that of the other. For…
Descriptors: Interrater Reliability, Test Bias, Multiple Choice Tests, Responses
Peer reviewed Peer reviewed
Direct linkDirect link
Jianbin Fu; Xuan Tan; Patrick C. Kyllonen – Journal of Educational Measurement, 2024
This paper presents the item and test information functions of the Rank two-parameter logistic models (Rank-2PLM) for items with two (pair) and three (triplet) statements in forced-choice questionnaires. The Rank-2PLM model for pairs is the MUPP-2PLM (Multi-Unidimensional Pairwise Preference) and, for triplets, is the Triplet-2PLM. Fisher's…
Descriptors: Questionnaires, Test Items, Item Response Theory, Models
Uk Hyun Cho – ProQuest LLC, 2024
The present study investigates the influence of multidimensionality on linking and equating in a unidimensional IRT. Two hypothetical multidimensional scenarios are explored under a nonequivalent group common-item equating design. The first scenario examines test forms designed to measure multiple constructs, while the second scenario examines a…
Descriptors: Item Response Theory, Classification, Correlation, Test Format
Peer reviewed Peer reviewed
Direct linkDirect link
Janet Mee; Ravi Pandian; Justin Wolczynski; Amy Morales; Miguel Paniagua; Polina Harik; Peter Baldwin; Brian E. Clauser – Advances in Health Sciences Education, 2024
Recent advances in automated scoring technology have made it practical to replace multiple-choice questions (MCQs) with short-answer questions (SAQs) in large-scale, high-stakes assessments. However, most previous research comparing these formats has used small examinee samples testing under low-stakes conditions. Additionally, previous studies…
Descriptors: Multiple Choice Tests, High Stakes Tests, Test Format, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Jonathan Hoseana; Andy Leonardo Louismono; Oriza Stepanus – International Journal of Mathematical Education in Science and Technology, 2025
We describe and evaluate a method to mitigate unwanted student collaborations in assessments, which we recently implemented in a second-year undergraduate mathematics module. The method requires a list of specific pairs of students to be prevented from collaborating, which we constructed based on the results of previous assessments. We converted…
Descriptors: Graphs, Color, College Mathematics, Undergraduate Students
Peer reviewed Peer reviewed
Direct linkDirect link
Ata Jahangir Moshayedi; Atanu Shuvam Roy; Zeashan Hameed Khan; Hong Lan; Habibollah Lotfi; Xiaohong Zhang – Education and Information Technologies, 2025
In this paper, a secure exam proctoring assistant 'EMTIHAN' (which means exam in Arabic/Persian/Urdu/Turkish languages) is developed to address concerns related to online exams for handwritten topics by allowing students to submit their answers online securely via their mobile devices. This system is designed with an aim to lessen the student's…
Descriptors: Computer Assisted Testing, Distance Education, MOOCs, Virtual Classrooms
Peer reviewed Peer reviewed
Direct linkDirect link
Harpreet Auby; Namrata Shivagunde; Vijeta Deshpande; Anna Rumshisky; Milo D. Koretsky – Journal of Engineering Education, 2025
Background: Analyzing student short-answer written justifications to conceptually challenging questions has proven helpful to understand student thinking and improve conceptual understanding. However, qualitative analyses are limited by the burden of analyzing large amounts of text. Purpose: We apply dense and sparse Large Language Models (LLMs)…
Descriptors: Student Evaluation, Thinking Skills, Test Format, Cognitive Processes
Peer reviewed Peer reviewed
Direct linkDirect link
Monica Casella; Pasquale Dolce; Michela Ponticorvo; Nicola Milano; Davide Marocco – Educational and Psychological Measurement, 2024
Short-form development is an important topic in psychometric research, which requires researchers to face methodological choices at different steps. The statistical techniques traditionally used for shortening tests, which belong to the so-called exploratory model, make assumptions not always verified in psychological data. This article proposes a…
Descriptors: Artificial Intelligence, Test Construction, Test Format, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Li Zhao; Junjie Peng; Shiqi Ke; Kang Lee – Educational Psychology Review, 2024
Unproctored and teacher-proctored exams have been widely used to prevent cheating at many universities worldwide. However, no empirical studies have directly compared their effectiveness in promoting academic integrity in actual exams. To address this significant gap, in four preregistered field studies, we examined the effectiveness of…
Descriptors: Supervision, Tests, Testing, Integrity
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  209