Publication Date
| In 2026 | 0 |
| Since 2025 | 0 |
| Since 2022 (last 5 years) | 2 |
| Since 2017 (last 10 years) | 13 |
| Since 2007 (last 20 years) | 40 |
Descriptor
| Evaluation Methods | 66 |
| Hypothesis Testing | 66 |
| Statistical Analysis | 66 |
| Research Methodology | 19 |
| Correlation | 15 |
| Educational Research | 13 |
| Comparative Analysis | 12 |
| Foreign Countries | 11 |
| Scores | 10 |
| Statistical Significance | 10 |
| Student Evaluation | 10 |
| More ▼ | |
Source
Author
Publication Type
Education Level
Location
| Finland | 2 |
| Nigeria | 2 |
| United Kingdom | 2 |
| Arkansas | 1 |
| Australia | 1 |
| Austria | 1 |
| Belgium | 1 |
| Canada | 1 |
| Denmark | 1 |
| Egypt | 1 |
| Florida | 1 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
| Defining Issues Test | 1 |
| Florida Comprehensive… | 1 |
| National Longitudinal… | 1 |
| Program for International… | 1 |
| Social Skills Rating System | 1 |
What Works Clearinghouse Rating
Tan, Teck Kiang – Practical Assessment, Research & Evaluation, 2023
Researchers often have hypotheses concerning the state of affairs in the population from which they sampled their data to compare group means. The classical frequentist approach provides one way of carrying out hypothesis testing using ANOVA to state the null hypothesis that there is no difference in the means and proceed with multiple comparisons…
Descriptors: Comparative Analysis, Hypothesis Testing, Statistical Analysis, Guidelines
Deke, John; Finucane, Mariel; Thal, Daniel – National Center for Education Evaluation and Regional Assistance, 2022
BASIE is a framework for interpreting impact estimates from evaluations. It is an alternative to null hypothesis significance testing. This guide walks researchers through the key steps of applying BASIE, including selecting prior evidence, reporting impact estimates, interpreting impact estimates, and conducting sensitivity analyses. The guide…
Descriptors: Bayesian Statistics, Educational Research, Data Interpretation, Hypothesis Testing
Yang, Shitao; Black, Ken – Teaching Statistics: An International Journal for Teachers, 2019
Summary Employing a Wald confidence interval to test hypotheses about population proportions could lead to an increase in Type I or Type II errors unless the hypothesized value, p0, is used in computing its standard error rather than the sample proportion. Whereas the Wald confidence interval to estimate a population proportion uses the sample…
Descriptors: Error Patterns, Evaluation Methods, Error of Measurement, Measurement Techniques
Porter, Kristin E. – Journal of Research on Educational Effectiveness, 2018
Researchers are often interested in testing the effectiveness of an intervention on multiple outcomes, for multiple subgroups, at multiple points in time, or across multiple treatment groups. The resulting multiplicity of statistical hypothesis tests can lead to spurious findings of effects. Multiple testing procedures (MTPs) are statistical…
Descriptors: Statistical Analysis, Program Effectiveness, Intervention, Hypothesis Testing
Hicks, Tyler; Rodríguez-Campos, Liliana; Choi, Jeong Hoon – American Journal of Evaluation, 2018
To begin statistical analysis, Bayesians quantify their confidence in modeling hypotheses with priors. A prior describes the probability of a certain modeling hypothesis apart from the data. Bayesians should be able to defend their choice of prior to a skeptical audience. Collaboration between evaluators and stakeholders could make their choices…
Descriptors: Bayesian Statistics, Evaluation Methods, Statistical Analysis, Hypothesis Testing
Porter, Kristin E. – Society for Research on Educational Effectiveness, 2016
In recent years, there has been increasing focus on the issue of multiple hypotheses testing in education evaluation studies. In these studies, researchers are typically interested in testing the effectiveness of an intervention on multiple outcomes, for multiple subgroups, at multiple points in time or across multiple treatment groups. When…
Descriptors: Hypothesis Testing, Intervention, Error Patterns, Evaluation Methods
Rajagopal, Prabha; Ravana, Sri Devi – Information Research: An International Electronic Journal, 2017
Introduction: The use of averaged topic-level scores can result in the loss of valuable data and can cause misinterpretation of the effectiveness of system performance. This study aims to use the scores of each document to evaluate document retrieval systems in a pairwise system evaluation. Method: The chosen evaluation metrics are document-level…
Descriptors: Information Retrieval, Documentation, Scores, Information Systems
Porter, Kristin E. – Grantee Submission, 2017
Researchers are often interested in testing the effectiveness of an intervention on multiple outcomes, for multiple subgroups, at multiple points in time, or across multiple treatment groups. The resulting multiplicity of statistical hypothesis tests can lead to spurious findings of effects. Multiple testing procedures (MTPs) are statistical…
Descriptors: Statistical Analysis, Program Effectiveness, Intervention, Hypothesis Testing
Suero, Manuel; Privado, Jesús; Botella, Juan – Psicologica: International Journal of Methodology and Experimental Psychology, 2017
A simulation study is presented to evaluate and compare three methods to estimate the variance of the estimates of the parameters d and "C" of the signal detection theory (SDT). Several methods have been proposed to calculate the variance of their estimators, "d'" and "c." Those methods have been mostly assessed by…
Descriptors: Evaluation Methods, Theories, Simulation, Statistical Analysis
Alzaid, Jawaher Mohammed – International Education Studies, 2017
This study aims at finding out the effect of peer assessment on the evaluation process of students. The hypothesis underlying this study is that assessment is an integral part of the learning process, which should play an important role in the educational model. The current study will emphasize the importance of using peer assessment as a tool to…
Descriptors: Foreign Countries, College Students, Peer Evaluation, Student Evaluation
Porter, Kristin E. – MDRC, 2016
In education research and in many other fields, researchers are often interested in testing the effectiveness of an intervention on multiple outcomes, for multiple subgroups, at multiple points in time, or across multiple treatment groups. The resulting multiplicity of statistical hypothesis tests can lead to spurious findings of effects. Multiple…
Descriptors: Statistical Analysis, Program Effectiveness, Intervention, Hypothesis Testing
Huggins-Manley, Anne Corinne – Educational and Psychological Measurement, 2017
This study defines subpopulation item parameter drift (SIPD) as a change in item parameters over time that is dependent on subpopulations of examinees, and hypothesizes that the presence of SIPD in anchor items is associated with bias and/or lack of invariance in three psychometric outcomes. Results show that SIPD in anchor items is associated…
Descriptors: Psychometrics, Test Items, Item Response Theory, Hypothesis Testing
Asrobi, Maman; Prasetyaningrum, Ari – English Language Teaching, 2017
This study was conducted in order to investigate the effectiveness of trait based assessment on teaching writing skill for EFL learners. Designed as pre-experimental study with one group pretest and posttest design, it examined 20 students of the second semester of English Department of "Hamzanwadi University" in the academic year…
Descriptors: English (Second Language), Second Language Learning, Second Language Instruction, Writing Skills
Dowling, Carey Bernini – International Journal for the Scholarship of Teaching and Learning, 2017
This study set out to replicate and extend research on students' reading compliance and examine the impact of daily quizzing methodology on students' reading compliance and retention. 98 students in two sections of Abnormal Psychology participated (mean age = 21.5, SD = 3.35; 72.4% Caucasian). Using a multiple baseline quasi-experimental design…
Descriptors: Undergraduate Students, Psychopathology, Evaluation Methods, Testing
Cyril, A. Vences; Jeyasekaran, D. – Journal on Educational Psychology, 2016
Continuous and Comprehensive Evaluation (CCE) refers to a system of school-based evaluation introduced by CBSE in all CBSE affiliated schools across the country to evaluate both scholastic and non-scholastic aspects of students' growth and development. Continuous and comprehensive evaluation is to evaluate every aspect of the child during their…
Descriptors: Student Evaluation, Attitude Measures, Student Attitudes, Foreign Countries

Peer reviewed
Direct link
