NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 12 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Yang Du; Susu Zhang – Journal of Educational and Behavioral Statistics, 2025
Item compromise has long posed challenges in educational measurement, jeopardizing both test validity and test security of continuous tests. Detecting compromised items is therefore crucial to address this concern. The present literature on compromised item detection reveals two notable gaps: First, the majority of existing methods are based upon…
Descriptors: Item Response Theory, Item Analysis, Bayesian Statistics, Educational Assessment
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Blair Lehman; Jesse R. Sparks; Jonathan Steinberg – ETS Research Report Series, 2024
Over the last 20 years, many methods have been proposed to use process data (e.g., response time) to detect changes in engagement during the test-taking process. However, many of these methods were developed and evaluated in highly similar testing contexts: 30 or more single-select multiple-choice items presented in a linear, fixed sequence in…
Descriptors: National Competency Tests, Secondary School Mathematics, Secondary School Students, Mathematics Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Kroehne, Ulf; Buerger, Sarah; Hahnel, Carolin; Goldhammer, Frank – Educational Measurement: Issues and Practice, 2019
For many years, reading comprehension in the Programme for International Student Assessment (PISA) was measured via paper-based assessment (PBA). In the 2015 cycle, computer-based assessment (CBA) was introduced, raising the question of whether central equivalence criteria required for a valid interpretation of the results are fulfilled. As an…
Descriptors: Reading Comprehension, Computer Assisted Testing, Achievement Tests, Foreign Countries
Wielicki, Tom – International Association for Development of the Information Society, 2016
This paper reports on longitudinal study regarding integrity of testing in an online format as used by e-learning platforms. Specifically, this study explains whether online testing, which implies an open book format is compromising integrity of assessment by encouraging cheating among students. Statistical experiment designed for this study…
Descriptors: Integrity, Online Courses, Statistical Surveys, Longitudinal Studies
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ali, Usama S.; Chang, Hua-Hua – ETS Research Report Series, 2014
Adaptive testing is advantageous in that it provides more efficient ability estimates with fewer items than linear testing does. Item-driven adaptive pretesting may also offer similar advantages, and verification of such a hypothesis about item calibration was the main objective of this study. A suitability index (SI) was introduced to adaptively…
Descriptors: Adaptive Testing, Simulation, Pretests Posttests, Test Items
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Becker, Kirk A.; Bergstrom, Betty A. – Practical Assessment, Research & Evaluation, 2013
The need for increased exam security, improved test formats, more flexible scheduling, better measurement, and more efficient administrative processes has caused testing agencies to consider converting the administration of their exams from paper-and-pencil to computer-based testing (CBT). Many decisions must be made in order to provide an optimal…
Descriptors: Testing, Models, Testing Programs, Program Administration
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Lin, Chuan-Ju – Journal of Technology, Learning, and Assessment, 2010
Assembling equivalent test forms with minimal test overlap across forms is important in ensuring test security. Chen and Lei (2009) suggested a exposure control technique to control test overlap-ordered item pooling on the fly based on the essence that test overlap rate--ordered item pooling for the first t examinees is a function of test overlap…
Descriptors: Test Length, Test Format, Evaluation Criteria, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Steinmetz, Jean-Paul; Brunner, Martin; Loarer, Even; Houssemand, Claude – Psychological Assessment, 2010
The Wisconsin Card Sorting Test (WCST) assesses executive and frontal lobe function and can be administered manually or by computer. Despite the widespread application of the 2 versions, the psychometric equivalence of their scores has rarely been evaluated and only a limited set of criteria has been considered. The present experimental study (N =…
Descriptors: Computer Assisted Testing, Psychometrics, Test Theory, Scores
Brodeur, Doris R. – Educational Technology, 1986
Reviews seven commercially produced test generator programs appropriate for use by classroom teachers or individual instructors and identifies item construction and test formatting features that facilitate test design and delivery. Test generator programs and their manufacturers are listed. (MBR)
Descriptors: Computer Assisted Testing, Computer Software, Costs, Evaluation Criteria
Vrasidas, Charalambos; Lantz, Chris – 1995
This paper describes a study in which a Picture Readability Index (PRI) was used to investigate initial and extended perceptions of photographs. Readability criteria for evaluating instructional text seems to have been in place for a long time, yet instructional visuals like photographs and illustrations have typically been subject to no such…
Descriptors: Adaptive Testing, Cognitive Processes, Computer Assisted Testing, Evaluation Criteria
Shannon, David M. – 1991
An overview is presented of methods of teacher assessment. The most heavily used methods of teacher assessment are standardized paper-and-pencil examinations and on-the-job ratings by supervisors. An alternative to the paper-and-pencil approach is explored, and preliminary evidence is presented in support of this assessment strategy.…
Descriptors: Computer Assisted Testing, Computer Simulation, Elementary School Teachers, Elementary Secondary Education
Anivan, Sarinee, Ed. – 1991
The selection of papers on language testing includes: "Language Testing in the 1990s: How Far Have We Come? How Much Further Have We To Go?" (J. Charles Alderson); "Current Research/Development in Language Testing" (John W. Oller, Jr.); "The Difficulties of Difficulty: Prompts in Writing Assessment" (Liz Hamp-Lyons,…
Descriptors: Communicative Competence (Languages), Comparative Analysis, Computer Assisted Testing, Cues