NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)0
Since 2006 (last 20 years)2
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 40 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Jordan, Sally – Computers & Education, 2012
Students were observed directly, in a usability laboratory, and indirectly, by means of an extensive evaluation of responses, as they attempted interactive computer-marked assessment questions that required free-text responses of up to 20 words and as they amended their responses after receiving feedback. This provided more general insight into…
Descriptors: Learner Engagement, Feedback (Response), Evaluation, Test Interpretation
National Council on Measurement in Education, 2012
Testing and data integrity on statewide assessments is defined as the establishment of a comprehensive set of policies and procedures for: (1) the proper preparation of students; (2) the management and administration of the test(s) that will lead to accurate and appropriate reporting of assessment results; and (3) maintaining the security of…
Descriptors: State Programs, Integrity, Testing, Test Preparation
Peer reviewed Peer reviewed
Harnisch, Delwyn L. – Journal of Educational Measurement, 1983
The Student-Problem (S-P) methodology is described using an example of 24 students on a test of 44 items. Information based on the students' test score and the modified caution index is put to diagnostic use. A modification of the S-P methodology is applied to domain-referenced testing. (Author/CM)
Descriptors: Academic Achievement, Educational Practices, Item Analysis, Responses
Carlson, Sybil B.; Ward, William C. – 1988
Issues concerning the cost and feasibility of using Formulating Hypotheses (FH) test item types for the Graduate Record Examinations have slowed research into their use. This project focused on two major issues that need to be addressed in considering FH items for operational use: the costs of scoring and the assignment of scores along a range of…
Descriptors: Adaptive Testing, Computer Assisted Testing, Costs, Pilot Projects
Bardovi-Harlig, Kathleen; Hartford, Beverly S. – 1993
A study compared the influence of two forms of discourse completion test (DCT) on the elicitation of rejection of advice. An open questionnaire providing scenarios alone was compared with a classic dialogue completion task in which a conversational turn is provided. The tasks were given to 32 graduate students, 19 native and 13 non-native…
Descriptors: Comparative Analysis, Dialogs (Language), Interpersonal Communication, Language Research
Lockhart, Kathleen A.; And Others – 1983
Three experiments were conducted, all employing undergraduates in college courses taught according to personalized system of instruction (PSI) principles. Experiment I examined retention as a function of the feedback delay interval in an introductory anthropology course using short-answer essay tests. Experiment II varied the feedback delay…
Descriptors: Cost Effectiveness, Feedback, Higher Education, Long Term Memory
Martinez, Michael E.; And Others – 1990
Large-scale testing is dominated by the multiple-choice question format. Widespread use of the format is due, in part, to the ease with which multiple-choice items can be scored automatically. This paper examines automatic scoring procedures for an alternative item type: figural response. Figural response items call for the completion or…
Descriptors: Automation, Computer Assisted Testing, Educational Technology, Multiple Choice Tests
Peer reviewed Peer reviewed
Gardner, Donald G.; Cummings, L. L.; Dunham, Randall B.; Pierce, Jon L. – Educational and Psychological Measurement, 1998
Whether traditional Likert-type focus of attention at work scales would outperform the one-item scales developed by D. Gardner and others (1989) was studied with responses of 492 automobile-services-club employees. Confirmatory factor analysis did not show either method to be better empirically. Situations in which the one-item scale might be…
Descriptors: Attention, Comparative Analysis, Employees, Likert Scales
Bensoussan, Marsha – 1993
Because testing comprehension of long (5- to 10-page) second-language texts is more complex than testing shorter texts, an alternative form of second language reading comprehension test is needed. In a guided summary completion test, students are presented with a greatly shortened version of the text (e.g., one-paragraph summary), containing…
Descriptors: English (Second Language), Foreign Countries, Higher Education, Language Tests
Arizona State Dept. of Education, Phoenix. – 1999
This guide is designed to give the student helpful information about Arizona's Instrument To Measure the Standards (AIMS) for high school. The guide provides examples of the type of questions students see on AIMS and explains how items are scored. It does not teach what the student needs to know to do well on AIMS, but it will help the student…
Descriptors: Academic Standards, High School Students, High Schools, Measurement Techniques
Arizona State Dept. of Education, Phoenix. – 2000
This addendum was created to reflect the recommendations of the Mathematics Task Force that met in Arizona to make recommendations to the state Board of Education. The Mathematics Task Force recommended that a core curriculum be identified in the Mathematics Proficiency Standards that would be required of all high school students beginning in the…
Descriptors: High School Students, High Schools, Mathematics, Measurement Techniques
Auchter, Joan E.; Stansfield, Charles – 1997
Is it possible to translate an existing test from the source language to a target language and then to empirically link the translated test to the source language standard? Some tests are amenable to translation of their content and some are not. Within tests, some items may be translatable and others may not. Some items may measure meaningful…
Descriptors: Bilingual Students, English, Equated Scores, Equivalency Tests
Martinez, Michael E.; Katz, Irvin R. – 1992
Contrasts between constructed response items and stem-equivalent multiple-choice counterparts typically have involved averaging item characteristics, and this aggregation has masked differences in statistical properties at the item level. Moreover, even aggregated format differences have not been explained in terms of differential cognitive…
Descriptors: Architecture, Cognitive Processes, Construct Validity, Constructed Response
Ohio State Univ., Columbus. Trade and Industrial Education Instructional Materials Lab. – 1978
The Ohio Vocational Achievement Tests are specially designed instruments for use by teachers, supervisors, and administrators to evaluate and diagnose vocational achievement for improving instruction in secondary vocational programs at the 11th and 12th grade levels. This guide explains the Ohio Vocational Achievement Tests and how they are used.…
Descriptors: Academic Achievement, Achievement Tests, High Schools, Scoring Formulas
Peer reviewed Peer reviewed
Frisbie, David A. – Educational Measurement: Issues and Practice, 1992
Literature related to the multiple true-false (MTF) item format is reviewed. Each answer cluster of a MTF item may have several true items and the correctness of each is judged independently. MTF tests appear efficient and reliable, although they are a bit harder than multiple choice items for examinees. (SLD)
Descriptors: Achievement Tests, Difficulty Level, Literature Reviews, Multiple Choice Tests
Previous Page | Next Page ยป
Pages: 1  |  2  |  3