NotesFAQContact Us
Collection
Advanced
Search Tips
Laws, Policies, & Programs
Race to the Top1
What Works Clearinghouse Rating
Showing 16 to 30 of 150 results Save | Export
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Xu, Peng; Desmarais, Michel C. – International Educational Data Mining Society, 2018
In most contexts of student skills assessment, whether the test material is administered by the teacher or within a learning environment, there is a strong incentive to minimize the number of questions or exercises administered in order to get an accurate assessment. This minimization objective can be framed as a Q-matrix design problem: given a…
Descriptors: Test Items, Accuracy, Test Construction, Skills
Peer reviewed Peer reviewed
Direct linkDirect link
Hill, Andrew P.; Donachie, Tracy – Journal of Psychoeducational Assessment, 2020
The measurement of perfectionistic cognitions has recently caused disagreement among researchers. Flett, Hewitt, Blankstein, and Gray proposed that perfectionistic cognitions are unidimensional. However, after re-examining the factor structure of the instrument used to measure perfectionistic automatic thoughts (Perfectionism Cognitions Inventory…
Descriptors: Factor Structure, Test Length, Cognitive Processes, Personality Traits
Peer reviewed Peer reviewed
Direct linkDirect link
Luo, Xiao; Wang, Xinrui – International Journal of Testing, 2019
This study introduced dynamic multistage testing (dy-MST) as an improvement to existing adaptive testing methods. dy-MST combines the advantages of computerized adaptive testing (CAT) and computerized adaptive multistage testing (ca-MST) to create a highly efficient and regulated adaptive testing method. In the test construction phase, multistage…
Descriptors: Adaptive Testing, Computer Assisted Testing, Test Construction, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Svetina, Dubravka; Liaw, Yuan-Ling; Rutkowski, Leslie; Rutkowski, David – Journal of Educational Measurement, 2019
This study investigates the effect of several design and administration choices on item exposure and person/item parameter recovery under a multistage test (MST) design. In a simulation study, we examine whether number-correct (NC) or item response theory (IRT) methods are differentially effective at routing students to the correct next stage(s)…
Descriptors: Measurement, Item Analysis, Test Construction, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Mameli, Consuelo; Passini, Stefano – Journal of Psychoeducational Assessment, 2019
The elusive character of student agency makes it a relevant construct to be investigated and measured. An initial effort in this direction was represented by the Agentic Engagement Scale, a five-item instrument designed to assess the degree to which students constructively contribute to the flow of the instructions they receive from the teacher.…
Descriptors: Measures (Individuals), Test Construction, Test Validity, Learner Engagement
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Manna, Venessa F.; Gu, Lixiong – ETS Research Report Series, 2019
When using the Rasch model, equating with a nonequivalent groups anchor test design is commonly achieved by adjustment of new form item difficulty using an additive equating constant. Using simulated 5-year data, this report compares 4 approaches to calculating the equating constants and the subsequent impact on equating results. The 4 approaches…
Descriptors: Item Response Theory, Test Items, Test Construction, Sample Size
Peer reviewed Peer reviewed
Direct linkDirect link
Xiao, Yang; Koenig, Kathleen; Han, Jing; Liu, Jing; Liu, Qiaoyi; Bao, Lei – Physical Review Physics Education Research, 2019
Standardized concept inventories (CIs) have been widely used in science, technology, engineering, and mathematics education for assessment of student learning. In practice, there have been concerns regarding the length of the test and possible test-retest memory effect. To address these issues, a recent study developed a method to split a CI into…
Descriptors: Scientific Concepts, Science Tests, Energy, Magnets
Peer reviewed Peer reviewed
Direct linkDirect link
Smith, William Zachary; Dickenson, Tammiee S.; Rogers, Bradley David – AERA Online Paper Repository, 2017
Questionnaire refinement and a process for selecting items for elimination are important tools for survey developers. One of the major obstacles in questionnaire refinement and elimination in surveys lies in one's ability to adequately and appropriately reconstruct a survey. Often times, surveys can be long and strenuous on the respondent,…
Descriptors: Surveys, Psychometrics, Test Construction, Test Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Liu, Ren; Huggins-Manley, Anne Corinne; Bradshaw, Laine – Educational and Psychological Measurement, 2017
There is an increasing demand for assessments that can provide more fine-grained information about examinees. In response to the demand, diagnostic measurement provides students with feedback on their strengths and weaknesses on specific skills by classifying them into mastery or nonmastery attribute categories. These attributes often form a…
Descriptors: Matrices, Classification, Accuracy, Diagnostic Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Mark L. Davison; David J. Weiss; Ozge Ersan; Joseph N. DeWeese; Gina Biancarosa; Patrick C. Kennedy – Grantee Submission, 2021
MOCCA is an online assessment of inferential reading comprehension for students in 3rd through 6th grades. It can be used to identify good readers and, for struggling readers, identify those who overly rely on either a Paraphrasing process or an Elaborating process when their comprehension is incorrect. Here a propensity to over-rely on…
Descriptors: Reading Tests, Computer Assisted Testing, Reading Comprehension, Elementary School Students
Peer reviewed Peer reviewed
Direct linkDirect link
Feher, Anita; Smith, Martin M.; Saklofske, Donald H.; Plouffe, Rachel A.; Wilson, Claire A.; Sherry, Simon B. – Journal of Psychoeducational Assessment, 2020
The Big Three Perfectionism Scale (BTPS) is a 45-item self-report measure of perfectionism with three overarching factors: rigid, self-critical, and narcissistic perfectionism. Our objective was to create a brief version of the BTPS, the Big Three Perfectionism Scale--Short Form (BTPS-SF). Sixteen items were selected, and confirmatory factor…
Descriptors: Personality Measures, Personality Traits, Test Construction, Measurement Techniques
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Sahin, Alper; Anil, Duygu – Educational Sciences: Theory and Practice, 2017
This study investigates the effects of sample size and test length on item-parameter estimation in test development utilizing three unidimensional dichotomous models of item response theory (IRT). For this purpose, a real language test comprised of 50 items was administered to 6,288 students. Data from this test was used to obtain data sets of…
Descriptors: Test Length, Sample Size, Item Response Theory, Test Construction
Peer reviewed Peer reviewed
Direct linkDirect link
Zaporozhets, Olga; Fox, Christine M.; Beltyukova, Svetlana A.; Laux, John M.; Piazza, Nick J.; Salyers, Kathleen – Measurement and Evaluation in Counseling and Development, 2015
This study was to develop a linear measure of change using University of Rhode Island Change Assessment items that represented Prochaska and DiClemente's theory. The resulting Toledo Measure of Change is short, is easy to use, and provides reliable scores for identification of individuals' stage of change and progression within that stage.
Descriptors: Item Response Theory, Change, Measures (Individuals), Test Construction
Jacob, Brian A. – Center on Children and Families at Brookings, 2016
Contrary to popular belief, modern cognitive assessments--including the new Common Core tests--produce test scores based on sophisticated statistical models rather than the simple percent of items a student answers correctly. While there are good reasons for this, it means that reported test scores depend on many decisions made by test designers,…
Descriptors: Scores, Common Core State Standards, Test Length, Test Content
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Li, Feifei – ETS Research Report Series, 2017
An information-correction method for testlet-based tests is introduced. This method takes advantage of both generalizability theory (GT) and item response theory (IRT). The measurement error for the examinee proficiency parameter is often underestimated when a unidimensional conditional-independence IRT model is specified for a testlet dataset. By…
Descriptors: Item Response Theory, Generalizability Theory, Tests, Error of Measurement
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10