NotesFAQContact Us
Collection
Advanced
Search Tips
Education Level
Audience
Researchers1
Laws, Policies, & Programs
Assessments and Surveys
Advanced Placement…1
What Works Clearinghouse Rating
Showing 1 to 15 of 25 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Johan Lyrvall; Zsuzsa Bakk; Jennifer Oser; Roberto Di Mari – Structural Equation Modeling: A Multidisciplinary Journal, 2024
We present a bias-adjusted three-step estimation approach for multilevel latent class models (LC) with covariates. The proposed approach involves (1) fitting a single-level measurement model while ignoring the multilevel structure, (2) assigning units to latent classes, and (3) fitting the multilevel model with the covariates while controlling for…
Descriptors: Hierarchical Linear Modeling, Statistical Bias, Error of Measurement, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
André Beauducel; Norbert Hilger; Tobias Kuhl – Educational and Psychological Measurement, 2024
Regression factor score predictors have the maximum factor score determinacy, that is, the maximum correlation with the corresponding factor, but they do not have the same inter-correlations as the factors. As it might be useful to compute factor score predictors that have the same inter-correlations as the factors, correlation-preserving factor…
Descriptors: Scores, Factor Analysis, Correlation, Predictor Variables
Peer reviewed Peer reviewed
Direct linkDirect link
Joakim Wallmark; James O. Ramsay; Juan Li; Marie Wiberg – Journal of Educational and Behavioral Statistics, 2024
Item response theory (IRT) models the relationship between the possible scores on a test item against a test taker's attainment of the latent trait that the item is intended to measure. In this study, we compare two models for tests with polytomously scored items: the optimal scoring (OS) model, a nonparametric IRT model based on the principles of…
Descriptors: Item Response Theory, Test Items, Models, Scoring
Peer reviewed Peer reviewed
Direct linkDirect link
Emma Somer; Carl Falk; Milica Miocevic – Structural Equation Modeling: A Multidisciplinary Journal, 2024
Factor Score Regression (FSR) is increasingly employed as an alternative to structural equation modeling (SEM) in small samples. Despite its popularity in psychology, the performance of FSR in multigroup models with small samples remains relatively unknown. The goal of this study was to examine the performance of FSR, namely Croon's correction and…
Descriptors: Scores, Structural Equation Models, Comparative Analysis, Sample Size
Peer reviewed Peer reviewed
Direct linkDirect link
Aidoo, Eric Nimako; Appiah, Simon K.; Boateng, Alexander – Journal of Experimental Education, 2021
This study investigated the small sample biasness of the ordered logit model parameters under multicollinearity using Monte Carlo simulation. The results showed that the level of biasness associated with the ordered logit model parameters consistently decreases for an increasing sample size while the distribution of the parameters becomes less…
Descriptors: Statistical Bias, Monte Carlo Methods, Simulation, Sample Size
Peer reviewed Peer reviewed
Direct linkDirect link
Kim, Stella Yun; Lee, Won-Chan – Applied Measurement in Education, 2023
This study evaluates various scoring methods including number-correct scoring, IRT theta scoring, and hybrid scoring in terms of scale-score stability over time. A simulation study was conducted to examine the relative performance of five scoring methods in terms of preserving the first two moments of scale scores for a population in a chain of…
Descriptors: Scoring, Comparative Analysis, Item Response Theory, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Koziol, Natalie A.; Goodrich, J. Marc; Yoon, HyeonJin – Educational and Psychological Measurement, 2022
Differential item functioning (DIF) is often used to examine validity evidence of alternate form test accommodations. Unfortunately, traditional approaches for evaluating DIF are prone to selection bias. This article proposes a novel DIF framework that capitalizes on regression discontinuity design analysis to control for selection bias. A…
Descriptors: Regression (Statistics), Item Analysis, Validity, Testing Accommodations
Peer reviewed Peer reviewed
Direct linkDirect link
Lübke, Karsten; Gehrke, Matthias; Horst, Jörg; Szepannek, Gero – Journal of Statistics Education, 2020
Basic knowledge of ideas of causal inference can help students to think beyond data, that is, to think more clearly about the data generating process. Especially for (maybe big) observational data, qualitative assumptions are important for the conclusions drawn and interpretation of the quantitative results. Concepts of causal inference can also…
Descriptors: Inferences, Simulation, Attribution Theory, Teaching Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Langan, Dean; Higgins, Julian P. T.; Jackson, Dan; Bowden, Jack; Veroniki, Areti Angeliki; Kontopantelis, Evangelos; Viechtbauer, Wolfgang; Simmonds, Mark – Research Synthesis Methods, 2019
Studies combined in a meta-analysis often have differences in their design and conduct that can lead to heterogeneous results. A random-effects model accounts for these differences in the underlying study effects, which includes a heterogeneity variance parameter. The DerSimonian-Laird method is often used to estimate the heterogeneity variance,…
Descriptors: Simulation, Meta Analysis, Health, Comparative Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Kontopantelis, Evangelos – Research Synthesis Methods, 2018
Background: Individual patient data (IPD) meta-analysis allows for the exploration of heterogeneity and can identify subgroups that most benefit from an intervention (or exposure), much more successfully than meta-analysis of aggregate data. One-stage or two-stage IPD meta-analysis is possible, with the former using mixed-effects regression models…
Descriptors: Patients, Medical Research, Meta Analysis, Intervention
Wang, Wei – ProQuest LLC, 2013
Mixed-format tests containing both multiple-choice (MC) items and constructed-response (CR) items are now widely used in many testing programs. Mixed-format tests often are considered to be superior to tests containing only MC items although the use of multiple item formats leads to measurement challenges in the context of equating conducted under…
Descriptors: Equated Scores, Test Format, Test Items, Test Length
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Ato, Manuel; Lopez, Juan Jose; Benavente, Ana – Psicologica: International Journal of Methodology and Experimental Psychology, 2011
A comparison between six rater agreement measures obtained using three different approaches was achieved by means of a simulation study. Rater coefficients suggested by Bennet's [sigma] (1954), Scott's [pi] (1955), Cohen's [kappa] (1960) and Gwet's [gamma] (2008) were selected to represent the classical, descriptive approach, [alpha] agreement…
Descriptors: Interrater Reliability, Measurement, Comparative Analysis, Statistical Analysis
Dong, Nianbo – Society for Research on Educational Effectiveness, 2011
The purpose of this study is through Monte Carlo simulation to compare several propensity score methods in approximating factorial experimental design and identify best approaches in reducing bias and mean square error of parameter estimates of the main and interaction effects of two factors. Previous studies focused more on unbiased estimates of…
Descriptors: Research Design, Probability, Monte Carlo Methods, Simulation
Peer reviewed Peer reviewed
Direct linkDirect link
Austin, Peter C. – Multivariate Behavioral Research, 2012
Researchers are increasingly using observational or nonrandomized data to estimate causal treatment effects. Essential to the production of high-quality evidence is the ability to reduce or minimize the confounding that frequently occurs in observational studies. When using the potential outcome framework to define causal treatment effects, one…
Descriptors: Computation, Regression (Statistics), Statistical Bias, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Cook, Thomas D.; Steiner, Peter M.; Pohl, Steffi – Multivariate Behavioral Research, 2009
This study uses within-study comparisons to assess the relative importance of covariate choice, unreliability in the measurement of these covariates, and whether regression or various forms of propensity score analysis are used to analyze the outcome data. Two of the within-study comparisons are of the four-arm type, and many more are of the…
Descriptors: Statistical Bias, Reliability, Data Analysis, Regression (Statistics)
Previous Page | Next Page »
Pages: 1  |  2