Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 0 |
| Since 2017 (last 10 years) | 1 |
| Since 2007 (last 20 years) | 8 |
Descriptor
| Comparative Analysis | 21 |
| Educational Assessment | 21 |
| Multiple Choice Tests | 21 |
| Foreign Countries | 7 |
| Performance Based Assessment | 7 |
| Elementary Secondary Education | 6 |
| Student Evaluation | 6 |
| Scoring | 5 |
| Test Construction | 5 |
| Test Format | 5 |
| Evaluation Methods | 4 |
| More ▼ | |
Source
Author
| Baron, Joan Boykoff | 1 |
| Barry, Carol | 1 |
| Cai, Li | 1 |
| Chang, Shu-Nu | 1 |
| Chen, Haiwen H. | 1 |
| Chiu, Mei-Hung | 1 |
| Crehan, Kevin | 1 |
| Delepine, Sidney G., III | 1 |
| Dost, Marcia A. | 1 |
| Dubois, Patrick J. | 1 |
| Edwards, K. Anthony | 1 |
| More ▼ | |
Publication Type
Education Level
| Higher Education | 2 |
| Postsecondary Education | 2 |
| Secondary Education | 2 |
| Elementary Education | 1 |
| Grade 6 | 1 |
| Grade 9 | 1 |
| High Schools | 1 |
| Intermediate Grades | 1 |
Audience
| Researchers | 1 |
Location
| Taiwan | 2 |
| United Kingdom (England) | 2 |
| Canada | 1 |
| Connecticut | 1 |
| France | 1 |
| Germany | 1 |
| Japan | 1 |
| Spain | 1 |
| United Kingdom (Wales) | 1 |
| United States | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| Program for International… | 1 |
What Works Clearinghouse Rating
Haberman, Shelby J.; Liu, Yang; Lee, Yi-Hsuan – ETS Research Report Series, 2019
Distractor analyses are routinely conducted in educational assessments with multiple-choice items. In this research report, we focus on three item response models for distractors: (a) the traditional nominal response (NR) model, (b) a combination of a two-parameter logistic model for item scores and a NR model for selections of incorrect…
Descriptors: Multiple Choice Tests, Scores, Test Reliability, High Stakes Tests
Falk, Carl F.; Cai, Li – Journal of Educational Measurement, 2016
We present a logistic function of a monotonic polynomial with a lower asymptote, allowing additional flexibility beyond the three-parameter logistic model. We develop a maximum marginal likelihood-based approach to estimate the item parameters. The new item response model is demonstrated on math assessment data from a state, and a computationally…
Descriptors: Item Response Theory, Guessing (Tests), Mathematics Tests, Simulation
Chen, Haiwen H.; von Davier, Matthias; Yamamoto, Kentaro; Kong, Nan – ETS Research Report Series, 2015
One major issue with large-scale assessments is that the respondents might give no responses to many items, resulting in less accurate estimations of both assessed abilities and item parameters. This report studies how the types of items affect the item-level nonresponse rates and how different methods of treating item-level nonresponses have an…
Descriptors: Achievement Tests, Foreign Countries, International Assessment, Secondary School Students
Paulhus, Delroy L.; Dubois, Patrick J. – Educational and Psychological Measurement, 2014
The overclaiming technique is a novel assessment procedure that uses signal detection analysis to generate indices of knowledge accuracy (OC-accuracy) and self-enhancement (OC-bias). The technique has previously shown robustness over varied knowledge domains as well as low reactivity across administration contexts. Here we compared the OC-accuracy…
Descriptors: Educational Assessment, Knowledge Level, Accuracy, Cognitive Ability
Delepine, Sidney G., III – ProQuest LLC, 2012
The purpose of this quantitative study is to compare a new assessment tool, the SkillsUSA Connect Assessment with the NOCTI assessment to determine which test results in more students achieving success. A quantitative study, designed to compare test scores of students taking the NOCTI assessment and new assessments from SkillsUSA, called the…
Descriptors: Educational Assessment, Academic Achievement, Scores, Comparative Analysis
Hsieh, Mingchuan – Language Assessment Quarterly, 2013
The Yes/No Angoff and Bookmark method for setting standards on educational assessment are currently two of the most popular standard-setting methods. However, there is no research into the comparability of these two methods in the context of language assessment. This study compared results from the Yes/No Angoff and Bookmark methods as applied to…
Descriptors: Standard Setting (Scoring), Comparative Analysis, Language Tests, Multiple Choice Tests
Kaliski, Pamela; Huff, Kristen; Barry, Carol – College Board, 2011
For educational achievement tests that employ multiple-choice (MC) items and aim to reliably classify students into performance categories, it is critical to design MC items that are capable of discriminating student performance according to the stated achievement levels. This is accomplished, in part, by clearly understanding how item design…
Descriptors: Alignment (Education), Academic Achievement, Expertise, Evaluative Thinking
Perdigones, Alicia; Garcia, Jose Luis; Valino, Vanesa; Raposo, Cecilia – Assessment & Evaluation in Higher Education, 2009
This work compares the results of three assessment systems used in two Spanish universities (the "Universidad Politecnica de Madrid" and the "Universidad Catolica de Avila"): the traditional system based on final examinations, continuous assessment with periodic tests and a proposed system (specially designed for heterogeneous…
Descriptors: Student Evaluation, Skill Development, Job Skills, Student Motivation
Peer reviewedEdwards, K. Anthony; Marshall, Carol – Teaching of Psychology, 1977
Describes a study of the accuracy of student responses on objective tests. Investigators examined the frequency of correctness on initial responses versus changed responses, and the relationship to degree of familiarity of the content. Results show that changing test answers tends to produce more right than wrong answers by more students.…
Descriptors: Comparative Analysis, Educational Assessment, Guessing (Tests), Higher Education
Baron, Joan Boykoff; And Others – 1981
Connecticut's experience with four different standard-setting methods regarding multiple choice proficiency tests is described. The methods include Angoff, Nedelsky, Borderline Group, and Contrasting Groups Methods. All Connecticut ninth graders were administered proficiency tests in reading, language arts, and mathematics. As soon as final test…
Descriptors: Academic Standards, Basic Skills, Comparative Analysis, Cutting Scores
Cognitive Complexity and the Comparability of Multiple-Choice and Constructed-Response Test Formats.
Peer reviewedHancock, Gregory R. – Journal of Experimental Education, 1994
To investigate the ability of multiple-choice tests to assess higher order thinking skills, examinations were constructed as half multiple choice and half constructed response. Results with 90 undergraduate and graduate students indicate that the 2 formats measure similar constructs at different levels of complexity. (SLD)
Descriptors: Cognitive Processes, Comparative Analysis, Constructed Response, Educational Assessment
Finch, F. L.; Dost, Marcia A. – 1992
Many state and local entities are developing and using performance assessment programs. Because these initiatives are so diverse, it is very difficult to understand what they are doing, or to compare them in any meaningful way. Multiple-choice tests are contrasted with performance assessments, and preliminary classifications are suggested to…
Descriptors: Alternative Assessment, Classification, Comparative Analysis, Constructed Response
Pollack, Judith M. – 1990
This paper summarizes an investigation of applications and issues in free response (FR) testing during 1989. It draws on ideas from the results of the National Educational Longitudinal Study 1988 (NELS:88) field test, a seminar series at the Educational Testing Service (ETS), working papers prepared for several FR testing applications, and…
Descriptors: Comparative Analysis, Costs, Educational Assessment, Elementary Secondary Education
Stecher, Brian – 1995
The resources necessary to create, administer, and score performance assessments in science were studied. RAND and the University of California, Santa Barbara (UCSB) designed performance tasks for science in grades five and six as part of a larger study of the feasibility of science performance assessment. Tasks were developed in pairs in task…
Descriptors: Classification, Comparative Analysis, Costs, Educational Assessment
Mehrens, William A. – 1991
Problems with performance assessment (PA) and multiple-choice tests (MCTs) are outlined, with reference to the literature on accountability. PA for individual teachers who should integrate their assessments with their instruction; PA as a supplement to more traditional examinations for licensure decisions; and some limited, experimental tryouts of…
Descriptors: Accountability, Comparative Analysis, Educational Assessment, Elementary Secondary Education
Previous Page | Next Page ยป
Pages: 1 | 2
Direct link
