Publication Date
| In 2026 | 0 |
| Since 2025 | 1 |
| Since 2022 (last 5 years) | 4 |
| Since 2017 (last 10 years) | 7 |
| Since 2007 (last 20 years) | 13 |
Descriptor
Source
Author
Publication Type
| Reports - Research | 25 |
| Journal Articles | 16 |
| Tests/Questionnaires | 3 |
| Speeches/Meeting Papers | 2 |
| Collected Works - Serials | 1 |
| Information Analyses | 1 |
Education Level
| Higher Education | 3 |
| Postsecondary Education | 3 |
| Preschool Education | 3 |
| Early Childhood Education | 2 |
| High Schools | 2 |
| Grade 1 | 1 |
| Kindergarten | 1 |
| Secondary Education | 1 |
Audience
| Researchers | 1 |
Laws, Policies, & Programs
Assessments and Surveys
| Graduate Management Admission… | 1 |
| Graduate Record Examinations | 1 |
| Raven Progressive Matrices | 1 |
What Works Clearinghouse Rating
Anirudhan Badrinath; Zachary Pardos – Journal of Educational Data Mining, 2025
Bayesian Knowledge Tracing (BKT) is a well-established model for formative assessment, with optimization typically using expectation maximization, conjugate gradient descent, or brute force search. However, one of the flaws of existing optimization techniques for BKT models is convergence to undesirable local minima that negatively impact…
Descriptors: Bayesian Statistics, Intelligent Tutoring Systems, Problem Solving, Audience Response Systems
Zachary J. Roman; Patrick Schmidt; Jason M. Miller; Holger Brandt – Structural Equation Modeling: A Multidisciplinary Journal, 2024
Careless and insufficient effort responding (C/IER) is a situation where participants respond to survey instruments without considering the item content. This phenomena adds noise to data leading to erroneous inference. There are multiple approaches to identifying and accounting for C/IER in survey settings, of these approaches the best performing…
Descriptors: Structural Equation Models, Bayesian Statistics, Response Style (Tests), Robustness (Statistics)
Liu, Tingting; Aryadoust, Vahid; Foo, Stacy – Language Testing, 2022
This study evaluated the validity of the Michigan English Test (MET) Listening Section by investigating its underlying factor structure and the replicability of its factor structure across multiple test forms. Data from 3255 test takers across four forms of the MET Listening Section were used. To investigate the factor structure, each form was…
Descriptors: Factor Structure, Language Tests, Second Language Learning, Second Language Instruction
Stoevenbelt, Andrea H.; Wicherts, Jelte M.; Flore, Paulette C.; Phillips, Lorraine A. T.; Pietschnig, Jakob; Verschuere, Bruno; Voracek, Martin; Schwabe, Inga – Educational and Psychological Measurement, 2023
When cognitive and educational tests are administered under time limits, tests may become speeded and this may affect the reliability and validity of the resulting test scores. Prior research has shown that time limits may create or enlarge gender gaps in cognitive and academic testing. On average, women complete fewer items than men when a test…
Descriptors: Timed Tests, Gender Differences, Item Response Theory, Correlation
Rodríguez-Vásquez, Flor Monserrat; Ariza-Hernandez, Francisco J. – EURASIA Journal of Mathematics, Science and Technology Education, 2021
The evaluation of learning in mathematics is a worldwide problem, therefore, new methods are required to assess the understanding of mathematical concepts. In this paper, we propose to use the Item Response Theory to analyze the understanding level of undergraduate students about the real function mathematical concept. The Bayesian approach was…
Descriptors: Bayesian Statistics, Mathematics Education, Item Response Theory, Undergraduate Students
Lúcio, Patrícia Silva; Vandekerckhove, Joachim; Polanczyk, Guilherme V.; Cogo-Moreira, Hugo – Journal of Psychoeducational Assessment, 2021
The present study compares the fit of two- and three-parameter logistic (2PL and 3PL) models of item response theory in the performance of preschool children on the Raven's Colored Progressive Matrices. The test of Raven is widely used for evaluating nonverbal intelligence of factor g. Studies comparing models with real data are scarce on the…
Descriptors: Guessing (Tests), Item Response Theory, Test Validity, Preschool Children
Kilic, Abdullah Faruk; Dogan, Nuri – International Journal of Assessment Tools in Education, 2021
Weighted least squares (WLS), weighted least squares mean-and-variance-adjusted (WLSMV), unweighted least squares mean-and-variance-adjusted (ULSMV), maximum likelihood (ML), robust maximum likelihood (MLR) and Bayesian estimation methods were compared in mixed item response type data via Monte Carlo simulation. The percentage of polytomous items,…
Descriptors: Factor Analysis, Computation, Least Squares Statistics, Maximum Likelihood Statistics
Mahmud, Jumailiyah; Sutikno, Muzayanah; Naga, Dali S. – Educational Research and Reviews, 2016
The aim of this study is to determine variance difference between maximum likelihood and expected A posteriori estimation methods viewed from number of test items of aptitude test. The variance presents an accuracy generated by both maximum likelihood and Bayes estimation methods. The test consists of three subtests, each with 40 multiple-choice…
Descriptors: Maximum Likelihood Statistics, Computation, Item Response Theory, Test Items
Scherer, Ronny; Meßinger-Koppelt, Jenny; Tiemann, Rüdiger – International Journal of STEM Education, 2014
Background: Complex problem-solving competence is regarded as a key construct in science education. But due to the necessity of using interactive and intransparent assessment procedures, appropriate measures of the construct are rare. This paper consequently presents the development and validation of a computer-based problem-solving environment,…
Descriptors: Computer Assisted Testing, Problem Solving, Chemistry, Science Tests
Young, Charles; Campbell, Megan – British Journal of Guidance & Counselling, 2014
This article provides GP-CORE norms for a South African university sample, which are compared to published data obtained from a United Kingdom university sample. The measure appears to be both reliable and valid for this multilingual and multicultural South African sample. The profiles of the psychological distress reported by white South African…
Descriptors: Foreign Countries, Well Being, Comparative Analysis, Psychological Needs
Zwick, Rebecca – ETS Research Report Series, 2012
Differential item functioning (DIF) analysis is a key component in the evaluation of the fairness and validity of educational tests. The goal of this project was to review the status of ETS DIF analysis procedures, focusing on three aspects: (a) the nature and stringency of the statistical rules used to flag items, (b) the minimum sample size…
Descriptors: Test Bias, Sample Size, Bayesian Statistics, Evaluation Methods
Sarmiento Campos, Jose A.; Squires, Jane; Ponte, Jaime – Early Child Development and Care, 2011
"A_Tempo" is a research project that is currently under development in Galicia, an autonomous community of Spain. Its main aim is to propose an effective universal screening procedure for early identification of developmental disorders in children from zero to three years of age who attend Galician pre-primary schools.…
Descriptors: Disability Identification, Foreign Countries, Screening Tests, Questionnaires
McDermott, Paul A.; Watkins, Marley W.; Rovine, Michael J.; Rikoon, Samuel H. – Journal of School Psychology, 2013
This article reports the development and evidence for validity and application of the Adjustment Scales for Early Transition in Schooling (ASETS). Based on primary analyses of data from the Head Start Impact Study, a nationally representative sample (N = 3077) of randomly selected children from low-income households is configured to inform…
Descriptors: Factor Analysis, Test Validity, Disadvantaged Youth, Social Adjustment
Peer reviewedHirsh, Hannah Rothstein; And Others – Personnel Psychology, 1986
Applied Schmidt-Hunter interactive validity generalization procedure to validity data for cognitive abilities tests for law enforcement occupations. Results indicated apparently lower validities and lesser generalizability for job criteria possibly due to low validity of the criterion and potential role of noncognitive factors. (Author/ABB)
Descriptors: Bayesian Statistics, Cognitive Tests, Generalizability Theory, Law Enforcement
Kingston, Neal M. – 1985
The incremental validity of the analytical measure of the revised Graduate Record Examination (GRE) General Test, for predicting first-year graduate grade-point average (GPA), was assessed using data submitted to the GRE Validity Study Service between March 1983 and November 1984. All selected students had data for the three General Test measures…
Descriptors: Aptitude Tests, Bayesian Statistics, College Entrance Examinations, Grade Point Average
Previous Page | Next Page »
Pages: 1 | 2
Direct link
