Publication Date
| In 2026 | 1 |
| Since 2025 | 104 |
| Since 2022 (last 5 years) | 556 |
| Since 2017 (last 10 years) | 12953 |
| Since 2007 (last 20 years) | 42088 |
Descriptor
| Statistical Analysis | 54731 |
| Foreign Countries | 19995 |
| Questionnaires | 10158 |
| Comparative Analysis | 9706 |
| Correlation | 9608 |
| Student Attitudes | 7428 |
| Teaching Methods | 6451 |
| Gender Differences | 6016 |
| College Students | 5858 |
| Qualitative Research | 5804 |
| Scores | 5577 |
| More ▼ | |
Source
Author
| Sinharay, Sandip | 37 |
| Thompson, Bruce | 37 |
| McNamara, Danielle S. | 36 |
| Raykov, Tenko | 36 |
| Smolkowski, Keith | 33 |
| Goldhaber, Dan | 31 |
| Vincent, Jack E. | 31 |
| Algina, James | 25 |
| Fien, Hank | 25 |
| Hwang, Gwo-Jen | 25 |
| Tsai, Chin-Chung | 25 |
| More ▼ | |
Publication Type
Education Level
Audience
| Researchers | 722 |
| Practitioners | 516 |
| Teachers | 412 |
| Administrators | 210 |
| Policymakers | 158 |
| Students | 79 |
| Media Staff | 35 |
| Counselors | 32 |
| Parents | 10 |
| Community | 9 |
| Support Staff | 4 |
| More ▼ | |
Location
| Turkey | 2624 |
| Australia | 1209 |
| Canada | 961 |
| California | 863 |
| China | 799 |
| Texas | 799 |
| Germany | 764 |
| Taiwan | 748 |
| Iran | 707 |
| United States | 703 |
| United Kingdom | 701 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
| Meets WWC Standards without Reservations | 50 |
| Meets WWC Standards with or without Reservations | 85 |
| Does not meet standards | 104 |
Peer reviewedKenneth Frank; Qinyun Lin; Spiro Maroulis; Shimeng Dai, Contributor; Nicole Jess, Contributor; Hung-Chang Lin, Contributor; Yuqing Liu, Contributor; Sarah Maestrales, Contributor; Ellen Searle, Contributor; Jordan Tait, Contributor – Grantee Submission, 2025
Sensitivity analyses can inform evidence-based education policy by quantifying the hypothetical conditions necessary to change an inference. Perhaps the most prevalent index used for sensitivity analyses is Oster's (2019) Coefficient of Proportionality (COP). Oster's COP leverages changes in estimated effects and R[superscript 2] when observed…
Descriptors: Statistical Analysis, Correlation, Predictor Variables, Inferences
Wendy Chan; Jimin Oh; Katherine J. Strickland – Society for Research on Educational Effectiveness, 2025
Background: The generalizability of a study refers to the extent to which the results and inferences from a sample apply to individuals in a larger target population of inference (Shadish et al., 2002). In practice, the strongest tool to facilitate generalizations is random or probability sampling, which is rare in educational studies (Olsen et…
Descriptors: Generalization, Sampling, Statistical Distributions, Statistical Analysis
Thorleif Lund – Scandinavian Journal of Educational Research, 2025
The purpose of this paper is to elaborate and comment on some themes related to construct validity, construct choice, and construct respecification in the Campbellian validity system. It is argued that Reichardt's [(2008, November). An alternative to the Campbellian conceptualization of validity. (Paper presentation). Evaluation 2008 Conference,…
Descriptors: Construct Validity, Statistical Analysis, Influences, Research Design
Daniel B. Wright – Open Education Studies, 2024
Pearson's correlation is widely used to test for an association between two variables and also forms the basis of several multivariate statistical procedures including many latent variable models. Spearman's [rho] is a popular alternative. These procedures are compared with ranking the data and then applying the inverse normal transformation, or…
Descriptors: Models, Simulation, Statistical Analysis, Correlation
Wen Luo; Haoran Li; Eunkyeng Baek; Chendong Li – Society for Research on Educational Effectiveness, 2025
Background/Context: Single-case experimental designs (SCEDs) play an important role in evaluating interventions in psychological, educational, and behavioral research. Unlike between-subjects designs, SCEDs involve a small number of cases whose responses to controlled experimental conditions are measured repeatedly over time. The evaluation of…
Descriptors: Effect Size, Research Design, Incidence, Intervention
Haeju Lee; Kyung Yong Kim – Journal of Educational Measurement, 2025
When no prior information of differential item functioning (DIF) exists for items in a test, either the rank-based or iterative purification procedure might be preferred. The rank-based purification selects anchor items based on a preliminary DIF test. For a preliminary DIF test, likelihood ratio test (LRT) based approaches (e.g.,…
Descriptors: Test Items, Equated Scores, Test Bias, Accuracy
Alan Huebner; Gustaf B. Skar; Mengchen Huang – Practical Assessment, Research & Evaluation, 2025
Generalizability theory is a modern and powerful framework for conducting reliability analyses. It is flexible to accommodate both random and fixed facets. However, there has been a relative scarcity in the practical literature on how to handle the fixed facet case. This article aims to provide practitioners a conceptual understanding and…
Descriptors: Generalizability Theory, Multivariate Analysis, Statistical Analysis, Writing Evaluation
Serpil Çelikten-Demirel; Aysenur Erdemir; Esra Oyar; Tuba Gündüz – International Journal of Assessment Tools in Education, 2025
It is an important point to test the homogeneity of variances in statistical methods such as the t-test or F-test used to make comparisons between groups. An erroneous decision regarding the homogeneity of variances will affect the test to be selected and thus lead to different results. For this reason, there are many tests for homogeneity of…
Descriptors: Statistical Analysis, Statistical Distributions, Sample Size, Error of Measurement
E. C. Hedberg; Larry V. Hedges – Evaluation Review, 2026
The difference in differences design is widely used to assess treatment effects in natural experiments or other situations where random assignment cannot, or is not, used (see, e.g., Angrist & Pischke, 2009). The researcher must make important decisions about which comparisons to make, the measurements to make, and perhaps the number of…
Descriptors: Statistical Analysis, Computation, Effect Size, Quasiexperimental Design
Lihan Chen; Milica Miocevic; Carl F. Falk – Structural Equation Modeling: A Multidisciplinary Journal, 2024
Data pooling is a powerful strategy in empirical research. However, combining multiple datasets often results in a large amount of missing data, as variables that are not present in some datasets effectively contain missing values for all participants in those datasets. Furthermore, data pooling typically leads to a mix of continuous and…
Descriptors: Simulation, Factor Analysis, Models, Statistical Analysis
Tom Benton – Practical Assessment, Research & Evaluation, 2025
This paper proposes an extension of linear equating that may be useful in one of two fairly common assessment scenarios. One is where different students have taken different combinations of test forms. This might occur, for example, where students have some free choice over the exam papers they take within a particular qualification. In this…
Descriptors: Equated Scores, Test Format, Test Items, Computation
Esfandiar Maasoumi; Le Wang; Daiqiang Zhang – Sociological Methods & Research, 2025
Current research on intergenerational mobility (IGM) is informed by "statistical" approaches based on log-level regressions, whose "economic" interpretations remain largely unknown. We reveal the subjective value-judgments in them: they are represented by weighted-sums (or aggregators) over heterogeneous groups, with…
Descriptors: Regression (Statistics), Social Mobility, Statistical Analysis, Income
Haoran Li; Chendong Li; Wen Luo; Eunkyeng Baek – Society for Research on Educational Effectiveness, 2025
Background/Context: Single-case experiment designs (SCEDs) are experimental designs in which a small number of cases are repeatedly measured over time, with manipulation of baseline and intervention phases. Because SCEDs often rely on direct behavioral observations, count data are common. To account for both the clustering and the non-normal…
Descriptors: Research Design, Effect Size, Statistical Analysis, Incidence
Paul A. Jewsbury; Daniel F. McCaffrey; Yue Jia; Eugenio J. Gonzalez – Journal of Educational Measurement, 2025
Large-scale survey assessments (LSAs) such as NAEP, TIMSS, PIRLS, IELS, and NAPLAN produce plausible values of student proficiency for estimating population statistics. Plausible values are imputed values for latent proficiency variables. While prominently used for LSAs, they are applicable to a wide range of latent variable modelling contexts…
Descriptors: Tests, Surveys, Monte Carlo Methods, Error of Measurement
Paul A. Jewsbury; Yue Jia; Eugenio J. Gonzalez – Large-scale Assessments in Education, 2024
Large-scale assessments are rich sources of data that can inform a diverse range of research questions related to educational policy and practice. For this reason, datasets from large-scale assessments are available to enable secondary analysts to replicate and extend published reports of assessment results. These datasets include multiple imputed…
Descriptors: Measurement, Data Analysis, Achievement, Statistical Analysis

Direct link
