Publication Date
| In 2026 | 0 |
| Since 2025 | 62 |
| Since 2022 (last 5 years) | 388 |
| Since 2017 (last 10 years) | 831 |
| Since 2007 (last 20 years) | 1345 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
| Practitioners | 195 |
| Teachers | 161 |
| Researchers | 93 |
| Administrators | 50 |
| Students | 34 |
| Policymakers | 15 |
| Parents | 12 |
| Counselors | 2 |
| Community | 1 |
| Media Staff | 1 |
| Support Staff | 1 |
| More ▼ | |
Location
| Canada | 63 |
| Turkey | 59 |
| Germany | 41 |
| United Kingdom | 37 |
| Australia | 36 |
| Japan | 35 |
| China | 33 |
| United States | 32 |
| California | 25 |
| Iran | 25 |
| United Kingdom (England) | 25 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Alemi, Minoo; Miraghaee, Apama – Journal on English Language Teaching, 2011
The present study was carried out to find out whether regular administration of cloze test improved the students' knowledge of grammar more than the multiple choice one. Subjects participating in this study were 84 Iranian pre-university students of Allameh-Gotb-e Ravandi University, aged between 18 and 35 and enrolled in a grammar course. To…
Descriptors: Foreign Countries, Comparative Analysis, Grammar, Knowledge Level
Jin, Yan – Journal of Pan-Pacific Association of Applied Linguistics, 2011
The College English Test (CET) is an English language test designed for educational purposes, administered on a very large scale, and used for making high-stakes decisions. This paper discusses the key issues facing the CET during the course of its development in the past two decades. It argues that the most fundamental and critical concerns of…
Descriptors: High Stakes Tests, Language Tests, Measures (Individuals), Graduates
Adebule, S. O. – Educational Research and Reviews, 2009
This study examined the reliability and difficult indices of Multiple Choice (MC) and True or False (TF) types of objective test items in a Mathematics Achievement Test (MAT). The instruments used were two variants- 50-items Mathematics achievement test based on the multiple choice and true or false test formats. A total of five hundred (500)…
Descriptors: Objective Tests, Mathematics Achievement, Achievement Tests, Test Reliability
Bennett, Talitha; De Bellis, David – Journal of Institutional Research, 2010
Like the vast majority of Australian universities, Flinders University (Flinders) collects feedback from students on the quality of teaching and learning through unit of study (topic) or classroom evaluations. Prior to 2009, survey instruments at Flinders were delivered via paper mode and in person to students in the classroom. In a drive for an…
Descriptors: Foreign Countries, Delivery Systems, Response Rates (Questionnaires), Educational Change
Lakin, Joni Marie – ProQuest LLC, 2010
Ability tests play an important role in the assessment programs of many schools. However, the inferences about ability made from such tests presume that students understand the tasks they are attempting. Task familiarity can vary by student as well as by format. By design, nonverbal reasoning tests use formats that are intended to be novel. The…
Descriptors: Test Items, Familiarity, Testing, Nonverbal Tests
Roberts, William L.; McKinley, Danette W.; Boulet, John R. – Advances in Health Sciences Education, 2010
Due to the high-stakes nature of medical exams it is prudent for test agencies to critically evaluate test data and control for potential threats to validity. For the typical multiple station performance assessments used in medicine, it may take time for examinees to become comfortable with the test format and administrative protocol. Since each…
Descriptors: Student Evaluation, Pretests Posttests, Licensing Examinations (Professions), Scores
Kim, Sooyeon; Walker, Michael E.; McHale, Frederick – ETS Research Report Series, 2008
This study examined variations of the nonequivalent-groups equating design for mixed-format tests--tests containing both multiple-choice (MC) and constructed-response (CR) items--to determine which design was most effective in producing equivalent scores across the two tests to be equated. Four linking designs were examined: (a) an anchor with…
Descriptors: Equated Scores, Test Format, Multiple Choice Tests, Responses
McGaw, Barry – Assessment in Education: Principles, Policy & Practice, 2008
In their reactions to my paper, the four authors provide comments that are illuminating and helpful for continuing discussions of the nature and utility of quantitative, comparative, international studies of educational achievement. In this response, I comment further on the issues of test characteristics, sample design, culture and causation.
Descriptors: Test Format, International Studies, Academic Achievement, Evaluation
van der Ark, L. Andries; Emons, Wilco H. M.; Sijtsma, Klaas – Journal of Educational Measurement, 2008
Two types of answer-copying statistics for detecting copiers in small-scale examinations are proposed. One statistic identifies the "copier-source" pair, and the other in addition suggests who is copier and who is source. Both types of statistics can be used when the examination has alternate test forms. A simulation study shows that the…
Descriptors: Cheating, Statistics, Test Format, Measures (Individuals)
Meyers, Jason L.; Murphy, Stephen; Goodman, Joshua; Turhan, Ahmet – Pearson, 2012
Operational testing programs employing item response theory (IRT) applications benefit from of the property of item parameter invariance whereby item parameter estimates obtained from one sample can be applied to other samples (when the underlying assumptions are satisfied). In theory, this feature allows for applications such as computer-adaptive…
Descriptors: Equated Scores, Test Items, Test Format, Item Response Theory
Romhild, Anja; Kenyon, Dorry; MacGregor, David – Language Assessment Quarterly, 2011
This study examined the role of domain-general and domain-specific linguistic knowledge in the assessment of academic English language proficiency using a latent variable modeling approach. The goal of the study was to examine if modeling of domain-specific variance results in improved model fit and well-defined latent factors. Analyses were…
Descriptors: Concept Formation, English (Second Language), Language Proficiency, Second Language Learning
Engaging Students and Evaluating Learning Progress Using Collaborative Exams in Introductory Courses
Eaton, Timothy T. – Journal of Geoscience Education, 2009
Collaborative exams, with subsections in which students have an opportunity to work with each other, are not yet widely used in introductory geoscience courses. This type of exam structure, with other participatory teaching strategies, was employed in two different courses, and results were found to provide a sensitive and revealing tool for…
Descriptors: Introductory Courses, Test Format, Student Evaluation, Evaluation Methods
Miyazaki, Kei; Hoshino, Takahiro; Mayekawa, Shin-ichi; Shigemasu, Kazuo – Psychometrika, 2009
This study proposes a new item parameter linking method for the common-item nonequivalent groups design in item response theory (IRT). Previous studies assumed that examinees are randomly assigned to either test form. However, examinees can frequently select their own test forms and tests often differ according to examinees' abilities. In such…
Descriptors: Test Format, Item Response Theory, Test Items, Test Bias
Lee, Kathryn S.; Osborne, Randall E.; Carpenter, David N. – Journal of Educational Computing Research, 2010
Prompted by a previous study investigating the effects of pacing on the academic testing performance of college students with AD/HD, we further explored our preliminary findings, which suggested that a computerized testing environment enhanced the testing performance of college students with AD/HD. We compared the effects of a computerized vs.…
Descriptors: College Students, Testing Accommodations, Testing, Attention Deficit Disorders
Kim, Sooyeon; Walker, Michael E.; McHale, Frederick – Journal of Educational Measurement, 2010
In this study we examined variations of the nonequivalent groups equating design for tests containing both multiple-choice (MC) and constructed-response (CR) items to determine which design was most effective in producing equivalent scores across the two tests to be equated. Using data from a large-scale exam, this study investigated the use of…
Descriptors: Measures (Individuals), Scoring, Equated Scores, Test Bias

Peer reviewed
Direct link
