NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Type
Information Analyses16
Reports - Evaluative16
Journal Articles8
Speeches/Meeting Papers3
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 16 results Save | Export
Emma Walland – Research Matters, 2024
GCSE examinations (taken by students aged 16 years in England) are not intended to be speeded (i.e. to be partly a test of how quickly students can answer questions). However, there has been little research exploring this. The aim of this research was to explore the speededness of past GCSE written examinations, using only the data from scored…
Descriptors: Educational Change, Test Items, Item Analysis, Scoring
Peer reviewed Peer reviewed
Direct linkDirect link
Cutumisu, Maria; Adams, Cathy; Lu, Chang – Journal of Science Education and Technology, 2019
Computational thinking (CT) is regarded as an essential twenty-first century competency and it is already embedded in K-12 curricula across the globe. However, research on assessing CT has lagged, with few assessments being implemented and validated. Moreover, there is a lack of systematic grouping of CT assessments. This scoping review examines…
Descriptors: Computation, Thinking Skills, 21st Century Skills, Elementary Secondary Education
Peer reviewed Peer reviewed
Direct linkDirect link
Wasanga, Paul; Somerset, Anthony – Assessment in Education: Principles, Policy & Practice, 2013
This paper starts with an analysis of the twin purposes of an examination reform programme originally launched in Kenya during the 1970s: first, to broaden the spectrum of cognitive skills being tested; and second, to set up a feedback system based on the performance profiles, providing schools and teachers with guidance as to how pedagogy and…
Descriptors: Foreign Countries, Exit Examinations, Change, Test Construction
Peer reviewed Peer reviewed
Haladyna, Thomas M.; Downing, Steven M. – Applied Measurement in Education, 1989
Results of 96 theoretical/empirical studies were reviewed to see if they support a taxonomy of 43 rules for writing multiple-choice test items. The taxonomy is the result of an analysis of 46 textbooks dealing with multiple-choice item writing. For nearly half of the rules, no research was found. (SLD)
Descriptors: Classification, Literature Reviews, Multiple Choice Tests, Test Construction
Hambleton, Ronald K.; Bollwark, John – 1991
The validity of results from international assessments depends on the correctness of the test translations. If the tests presented in one language are more or less difficult because of the manner in which they are translated, the validity of any interpretation of the results can be questioned. Many test translation methods exist in the literature,…
Descriptors: Cultural Differences, Educational Assessment, English, Foreign Countries
Peer reviewed Peer reviewed
Knowles, Susan L.; Welch, Cynthia A. – Educational and Psychological Measurement, 1992
A meta-analysis of the difficulty and discrimination of the "none-of-the-above" (NOTA) test option was conducted with 12 articles (20 effect sizes) for difficulty and 7 studies (11 effect sizes) for discrimination. Findings indicate that using the NOTA option does not result in items of lesser quality. (SLD)
Descriptors: Difficulty Level, Effect Size, Meta Analysis, Multiple Choice Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Read, John – International Journal of English Studies, 2007
This paper surveys some current developments in second language vocabulary assessment, with particular attention to the ways in which computer corpora can provide better quality information about the frequency of words and how they are used in specific contexts. The relative merits of different word lists are discussed, including the Academic Word…
Descriptors: Second Language Learning, Second Language Programs, Vocabulary, Educational Development
Rosner, Frieda C.; Weber, Wilford A. – 1982
A review of the National Teacher Examinations (NTE) has focused on the Commons Examinations component. The Commons, now named the National Teacher Examinations Core Battery, tests general knowledge, communication skills, and professional knowledge in three separate tests. Users may select portions of the tests that best suit their needs at various…
Descriptors: Classroom Techniques, Higher Education, Program Evaluation, Standardized Tests
Peer reviewed Peer reviewed
Frisbie, David A. – Educational Measurement: Issues and Practice, 1992
Literature related to the multiple true-false (MTF) item format is reviewed. Each answer cluster of a MTF item may have several true items and the correctness of each is judged independently. MTF tests appear efficient and reliable, although they are a bit harder than multiple choice items for examinees. (SLD)
Descriptors: Achievement Tests, Difficulty Level, Literature Reviews, Multiple Choice Tests
Wainer, Howard; Thissen, David – 1994
When an examination consists in whole or part of constructed response test items, it is common practice to allow the examinee to choose a subset of the constructed response questions from a larger pool. It is sometimes argued that, if choice were not allowed, the limitations on domain coverage forced by the small number of items might unfairly…
Descriptors: Constructed Response, Difficulty Level, Educational Testing, Equated Scores
Colton, Dean A. – 1993
Tables of specifications are used to guide test developers in sampling items and maintaining consistency from form to form. This paper is a generalizability study of the American College Testing Program (ACT) Achievement Program Mathematics Test (AAP), with the content areas of the table of specifications representing multiple dependent variables.…
Descriptors: Achievement Tests, Difficulty Level, Error of Measurement, Generalizability Theory
Stansfield, Charles W. – 1990
The IDEA Oral Language Proficiency Test (IPT II), an individually-administered measure of speaking and listening proficiency in English as a Second Language designed for secondary school students, is described and discussed. The test consists of 91 items and requires 5-25 minutes to administer. Raw scores are converted to one of seven proficiency…
Descriptors: Classification, English (Second Language), Language Proficiency, Language Tests
Murray, Joel R. – 2001
This paper aims to provide practical advice for creating a placement test for English-as-a-Second-Language (ESL) or English-as-a-foreign-language (EFL) instruction. Three forms of concrete assistance are provided: a detailed literature review; detailed steps focusing on the creation of placement tests; and a set of recommendations focusing on…
Descriptors: English (Second Language), Examiners, Factor Analysis, Literature Reviews
Peer reviewed Peer reviewed
Abrami, Philip C.; d'Apollonia, Sylvia – New Directions for Teaching and Learning, 1990
Multidimensional rating forms used for college student evaluation of faculty elicit information that can be used to evaluate the dimensions of instruction either separately or globally. Portions of the information may be useful for formative evaluation, but global items are best for making personnel decisions. (MSE)
Descriptors: College Faculty, Decision Making, Educational Improvement, Employment Practices
Perez, Kristina M. – 1996
The KeyMath Revised is a power test that measures the understanding and application of mathematics skills and concepts. It is individually administered and is intended for students from kindergarten through the ninth grade to determine student mastery of mathematics concepts. The revised version is designed to be user-friendly for the student and…
Descriptors: Comprehension, Curriculum Development, Diagnostic Tests, Educational Diagnosis
Previous Page | Next Page ยป
Pages: 1  |  2