NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1 to 15 of 289 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Jianbin Fu; Xuan Tan; Patrick C. Kyllonen – Journal of Educational Measurement, 2024
This paper presents the item and test information functions of the Rank two-parameter logistic models (Rank-2PLM) for items with two (pair) and three (triplet) statements in forced-choice questionnaires. The Rank-2PLM model for pairs is the MUPP-2PLM (Multi-Unidimensional Pairwise Preference) and, for triplets, is the Triplet-2PLM. Fisher's…
Descriptors: Questionnaires, Test Items, Item Response Theory, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Jiawei Xiong; George Engelhard; Allan S. Cohen – Measurement: Interdisciplinary Research and Perspectives, 2025
It is common to find mixed-format data results from the use of both multiple-choice (MC) and constructed-response (CR) questions on assessments. Dealing with these mixed response types involves understanding what the assessment is measuring, and the use of suitable measurement models to estimate latent abilities. Past research in educational…
Descriptors: Responses, Test Items, Test Format, Grade 8
Peer reviewed Peer reviewed
Direct linkDirect link
Choe, Edison M.; Han, Kyung T. – Journal of Educational Measurement, 2022
In operational testing, item response theory (IRT) models for dichotomous responses are popular for measuring a single latent construct [theta], such as cognitive ability in a content domain. Estimates of [theta], also called IRT scores or [theta hat], can be computed using estimators based on the likelihood function, such as maximum likelihood…
Descriptors: Scores, Item Response Theory, Test Items, Test Format
Peer reviewed Peer reviewed
Direct linkDirect link
van der Linden, Wim J. – Journal of Educational and Behavioral Statistics, 2022
The current literature on test equating generally defines it as the process necessary to obtain score comparability between different test forms. The definition is in contrast with Lord's foundational paper which viewed equating as the process required to obtain comparability of measurement scale between forms. The distinction between the notions…
Descriptors: Equated Scores, Test Items, Scores, Probability
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Mustafa Ilhan; Nese Güler; Gülsen Tasdelen Teker; Ömer Ergenekon – International Journal of Assessment Tools in Education, 2024
This study aimed to examine the effects of reverse items created with different strategies on psychometric properties and respondents' scale scores. To this end, three versions of a 10-item scale in the research were developed: 10 positive items were integrated in the first form (Form-P) and five positive and five reverse items in the other two…
Descriptors: Test Items, Psychometrics, Scores, Measures (Individuals)
Peer reviewed Peer reviewed
Direct linkDirect link
Jianbin Fu; Patrick C. Kyllonen; Xuan Tan – Measurement: Interdisciplinary Research and Perspectives, 2024
Users of forced-choice questionnaires (FCQs) to measure personality commonly assume statement parameter invariance across contexts -- between Likert and forced-choice (FC) items and between different FC items that share a common statement. In this paper, an empirical study was designed to check these two assumptions for an FCQ assessment measuring…
Descriptors: Measurement Techniques, Questionnaires, Personality Measures, Interpersonal Competence
Peer reviewed Peer reviewed
Direct linkDirect link
Hye-won Lee; Andrew Mullooly; Amy Devine; Evelina Galaczi – Applied Linguistics, 2024
In the assessment of second language oral communication, the video-call speaking test has received increasing attention as a test method with higher practicality than its in-person counterpart, but still with broad coverage of the test construct. Previous studies into video-call assessment have focussed on the individual (as opposed to paired or…
Descriptors: Oral Language, Language Skills, Speech Communication, Interaction Process Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Green, Theresa; Goodridge, Wade H.; Anderson, Jon; Davishahl, Eric; Kane, Daniel – International Education Studies, 2023
The purpose of this study was to examine any differences in test scores between three different online versions of the Mental Cutting Test (MCT). The MCT was developed to quantify a rotational and proportion construct of spatial ability and has been used extensively to assess spatial ability. This test was developed in 1938 as a paper-and-pencil…
Descriptors: Spatial Ability, Measures (Individuals), Computer Assisted Testing, Test Format
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Robert N. Prince – Numeracy, 2025
One of the effects of the COVID-19 pandemic was the rapid shift to replacing traditional, paper-based tests with their computer-based counterparts. In many cases, these new modes of delivering tests will remain in place for the foreseeable future. In South Africa, the National Benchmark Quantitative Literacy (QL) test was impelled to make this…
Descriptors: Benchmarking, Numeracy, Multiple Literacies, Paper and Pencil Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Sheri Bayley – Teaching and Learning in Communication Sciences & Disorders, 2024
The purpose of this study was to explore student performance, self-ratings of learning and preference, and student comments on a variety of reading quiz formats in a first semester speech-language pathology graduate course. Students from two cohorts (n = 34) completed four types of quizzes: closed-book, open-book, open-note, and collaborative…
Descriptors: Reading Instruction, Tests, Graduate Students, Courses
Santi Lestari – Research Matters, 2024
Despite the increasing ubiquity of computer-based tests, many general qualifications examinations remain in a paper-based mode. Insufficient and unequal digital provision across schools is often identified as a major barrier to a full adoption of computer-based exams for general qualifications. One way to overcome this barrier is a gradual…
Descriptors: Keyboarding (Data Entry), Handwriting, Test Format, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Thuy Ho Hoang Nguyen; Bao Trang Thi Nguyen; Giang Thi Linh Hoang; Nhung Thi Hong Pham; Tu Thi Cam Dang – Language Testing in Asia, 2024
The present study explored the comparability in performance scores between the computer-delivered and face-to-face modes for the two speaking tests in the Vietnamese Standardized Test of English Proficiency (VSTEP) (the VSTEP.2 and VSTEP.3-5 Speaking tests) according to Vietnam's Six-Level Foreign Language Proficiency Framework (VNFLPF) and test…
Descriptors: Test Format, Computer Assisted Testing, Student Attitudes, Language Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Jones, Paul; Tong, Ye; Liu, Jinghua; Borglum, Joshua; Primoli, Vince – Journal of Educational Measurement, 2022
This article studied two methods to detect mode effects in two credentialing exams. In Study 1, we used a "modal scale comparison approach," where the same pool of items was calibrated separately, without transformation, within two TC cohorts (TC1 and TC2) and one OP cohort (OP1) matched on their pool-based scale score distributions. The…
Descriptors: Scores, Credentials, Licensing Examinations (Professions), Computer Assisted Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Baldwin, Peter; Clauser, Brian E. – Journal of Educational Measurement, 2022
While score comparability across test forms typically relies on common (or randomly equivalent) examinees or items, innovations in item formats, test delivery, and efforts to extend the range of score interpretation may require a special data collection before examinees or items can be used in this way--or may be incompatible with common examinee…
Descriptors: Scoring, Testing, Test Items, Test Format
Peer reviewed Peer reviewed
Direct linkDirect link
Sen, Sedat – Creativity Research Journal, 2022
The purpose of this study was to estimate the overall reliability values for the scores produced by Runco Ideational Behavior Scale (RIBS) and explore the variability of RIBS score reliability across studies. To achieve this, a reliability generalization meta-analysis was carried out using the 86 Cronbach's alpha estimates obtained from 77 studies…
Descriptors: Generalization, Creativity, Meta Analysis, Higher Education
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  20