NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 646 to 660 of 1,187 results Save | Export
Hamzah, Mohd Sahandri Gani; Abdullah, Saifuddin Kumar – Online Submission, 2011
The evaluation of learning is a systematic process involving testing, measuring and evaluation. In the testing step, a teacher needs to choose the best instrument that can test the minds of students. Testing will produce scores or marks with many variations either in homogeneous or heterogeneous forms that will be used to categorize the scores…
Descriptors: Test Items, Item Analysis, Difficulty Level, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Bernholt, Sascha; Parchmann, Ilka – Chemistry Education Research and Practice, 2011
Current reforms in the education policy of various countries are intended to produce a paradigm shift in the educational system towards an outcome orientation. After implementing educational standards as normative objectives, the development of test procedures that adequately reflect these targets and standards is a central problem. This paper…
Descriptors: Science Achievement, Chemistry, Knowledge Level, Science Instruction
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Styles, Irene; Wildy, Helen; Pepper, Vivienne; Faulkner, Joanne; Berman, Ye'Elah – International Research in Early Childhood Education, 2014
The assessment of literacy and numeracy skills of students as they enter school for the first time is not yet established nation-wide in Australia. However, a large proportion of primary schools have chosen to assess their starting students on the Performance Indicators in Primary Schools-Baseline Assessment (PIPS-BLA). This series of three…
Descriptors: Foreign Countries, Indigenous Knowledge, Performance Based Assessment, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Hough, Heather J.; Kerbow, David; Bryk, Anthony; Pinnell, Gay Su; Rodgers, Emily; Dexter, Emily; Hung, Carrie; Scharer, Patricia L.; Fountas, Irene – School Effectiveness and School Improvement, 2013
In this paper, we report on 2 studies developing, testing, and using an observation tool for measuring primary literacy instruction, the Developing Language and Literacy Teaching (DLLT) rubrics. In Study 1 (an instrumentation study), we show that the DLLT has a high level of internal consistency, that there are high levels of inter-rater…
Descriptors: Literacy Education, Teacher Evaluation, Observation, Scoring Rubrics
Peer reviewed Peer reviewed
Direct linkDirect link
Rufino, Katrina A.; Boccaccini, Marcus T.; Guy, Laura S. – Assessment, 2011
Although reliability is essential to validity, most research on violence risk assessment tools has paid little attention to strategies for improving rater agreement. The authors evaluated the degree to which perceived subjectivity in scoring guidelines for items from two measures--the Psychopathy Checklist-Revised (PCL-R) and the Historical,…
Descriptors: Risk Management, Predictive Validity, Interrater Reliability, Scoring
OECD Publishing, 2014
The "PISA 2012 Technical Report" describes the methodology underlying the PISA 2012 survey, which tested 15-year-olds' competencies in mathematics, reading and science and, in some countries, problem solving and financial literacy. It examines the design and implementation of the project at a level of detail that allows researchers to…
Descriptors: International Assessment, Secondary School Students, Foreign Countries, Achievement Tests
Tortop, Hasan Said – Online Submission, 2012
A field trip is an excursion by group of students with teachers to a place away from classroom such as natural field, science center, and zoo. So, it is an important tool for renewable energy education. This study was carried out to develop a new scale for measuring teacher attitudes towards the field trip. Teacher attitude scale towards the field…
Descriptors: Test Construction, Field Trips, Teacher Attitudes, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Hasson, Natalie; Dodd, Barbara; Botting, Nicola – International Journal of Language & Communication Disorders, 2012
Background: Sentence construction and syntactic organization are known to be poor in children with specific language impairments (SLI), but little is known about the way in which children with SLI approach language tasks, and static standardized tests contribute little to the differentiation of skills within the population of children with…
Descriptors: Alternative Assessment, Sentence Structure, Syntax, Language Processing
Peer reviewed Peer reviewed
Direct linkDirect link
Day, James; Bonn, Doug – Physical Review Special Topics - Physics Education Research, 2011
The Concise Data Processing Assessment (CDPA) was developed to probe student abilities related to the nature of measurement and uncertainty and to handling data. The diagnostic is a ten question, multiple-choice test that can be used as both a pre-test and post-test. A key component of the development process was interviews with students, which…
Descriptors: Multiple Choice Tests, Test Reliability, Physics, Item Analysis
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Rogers, Daniel T. – International Journal for the Scholarship of Teaching and Learning, 2012
Despite potential applications to educational contexts, the working alliance concept has largely been confined to psychotherapy intervention research. Some have explored theoretically related concepts (e.g., immediacy, rapport), but no measure currently exists of the working alliance between a teacher and student within an academic course. The aim…
Descriptors: Test Construction, Test Validity, Test Reliability, Psychometrics
Irvin, P. Shawn; Alonzo, Julie; Lai, Cheng-Fei; Park, Bitnara Jasmine; Tindal, Gerald – Behavioral Research and Teaching, 2012
In this technical report, we present the results of a reliability study of the seventh-grade multiple choice reading comprehension measures available on the easyCBM learning system conducted in the spring of 2011. Analyses include split-half reliability, alternate form reliability, person and item reliability as derived from Rasch analysis,…
Descriptors: Reading Comprehension, Testing Programs, Statistical Analysis, Grade 7
Peer reviewed Peer reviewed
Direct linkDirect link
Lane, Kathleen Lynne; Menzies, Holly M.; Oakes, Wendy P.; Lambert, Warren; Cox, Meredith; Hankins, Katy – Behavioral Disorders, 2012
We report findings of two studies, one conducted in a rural school district (N = 982) and a second conducted in an urban district (N = 1,079), offering additional evidence of the reliability and validity of a revised instrument, the Student Risk Screening Scale-Internalizing and Externalizing (SRSS-IE), to accurately detect internalizing and…
Descriptors: Test Validity, Behavior Disorders, Measures (Individuals), Screening Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Bristow, M.; Erkorkmaz, K.; Huissoon, J. P.; Jeon, Soo; Owen, W. S.; Waslander, S. L.; Stubley, G. D. – IEEE Transactions on Education, 2012
Any meaningful initiative to improve the teaching and learning in introductory control systems courses needs a clear test of student conceptual understanding to determine the effectiveness of proposed methods and activities. The authors propose a control systems concept inventory. Development of the inventory was collaborative and iterative. The…
Descriptors: Diagnostic Tests, Concept Formation, Undergraduate Students, Engineering Education
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Sahin, Ismail – Turkish Online Journal of Educational Technology - TOJET, 2011
The purpose of this study is to develop a survey of technological pedagogical and content knowledge (TPACK). The survey consists of seven subscales forming the TPACK model: 1) technology knowledge (TK), 2) pedagogy knowledge (PK), 3) content knowledge (CK), 4) technological pedagogical knowledge (TPK), 5) technological content knowledge (TCK), 6)…
Descriptors: Preservice Teachers, Test Validity, Pedagogical Content Knowledge, Surveys
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Lee, Yi-Hsuan; Zhang, Jinming – ETS Research Report Series, 2010
This report examines the consequences of differential item functioning (DIF) using simulated data. Its impact on total score, item response theory (IRT) ability estimate, and test reliability was evaluated in various testing scenarios created by manipulating the following four factors: test length, percentage of DIF items per form, sample sizes of…
Descriptors: Test Bias, Item Response Theory, Test Items, Scores
Pages: 1  |  ...  |  40  |  41  |  42  |  43  |  44  |  45  |  46  |  47  |  48  |  ...  |  80