Publication Date
| In 2026 | 0 |
| Since 2025 | 2 |
| Since 2022 (last 5 years) | 6 |
| Since 2017 (last 10 years) | 8 |
| Since 2007 (last 20 years) | 10 |
Descriptor
| Factor Analysis | 15 |
| Item Analysis | 15 |
| Multiple Choice Tests | 15 |
| Test Items | 8 |
| Test Reliability | 8 |
| Test Construction | 6 |
| Item Response Theory | 5 |
| Comparative Analysis | 4 |
| Foreign Countries | 4 |
| Statistical Analysis | 4 |
| Test Validity | 4 |
| More ▼ | |
Source
Author
| Andrew Gardiner | 1 |
| Apantee Poonputta | 1 |
| Baghaei, Purya | 1 |
| Bello, Samira Abdullahi | 1 |
| Bichi, Ado Abdu | 1 |
| Bulut, Okan | 1 |
| Chu, Wei | 1 |
| Darren J. Shaw | 1 |
| Dedi Kuswandi | 1 |
| Dourakhshan, Alireza | 1 |
| Dywel, Malwina | 1 |
| More ▼ | |
Publication Type
| Reports - Research | 13 |
| Journal Articles | 8 |
| Speeches/Meeting Papers | 3 |
Education Level
| Higher Education | 4 |
| Postsecondary Education | 4 |
| Secondary Education | 3 |
| High Schools | 2 |
| Elementary Education | 1 |
| Grade 12 | 1 |
Audience
| Researchers | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Thayaamol Upapong; Apantee Poonputta – Educational Process: International Journal, 2025
Background/purpose: The purposes of this research are to develop a reliable and valid assessment tool for measuring systems thinking skills in upper primary students in Thailand and to establish a normative criterion for evaluating their systems thinking abilities based on educational standards. Materials/methods: The study followed a three-phase…
Descriptors: Thinking Skills, Elementary School Students, Measures (Individuals), Foreign Countries
Testing Anatomy: Dissecting Spatial and Non-Spatial Knowledge in Multiple-Choice Question Assessment
Julie Dickson; Darren J. Shaw; Andrew Gardiner; Susan Rhind – Anatomical Sciences Education, 2024
Limited research has been conducted on the spatial ability of veterinary students and how this is evaluated within anatomy assessments. This study describes the creation and evaluation of a split design multiple-choice question (MCQ) assessment (totaling 30 questions divided into 15 non-spatial MCQs and 15 spatial MCQs). Two cohorts were tested,…
Descriptors: Anatomy, Spatial Ability, Multiple Choice Tests, Factor Analysis
Güntay Tasçi – Science Insights Education Frontiers, 2024
The present study has aimed to develop and validate a protein concept inventory (PCI) consisting of 25 multiple-choice (MC) questions to assess students' understanding of protein, which is a fundamental concept across different biology disciplines. The development process of the PCI involved a literature review to identify protein-related content,…
Descriptors: Science Instruction, Science Tests, Multiple Choice Tests, Biology
Nurussaniah Nurussaniah; Punaji Setyosari; Dedi Kuswandi; Saida Ulfa – Journal of Baltic Science Education, 2025
The accurate assessment of analytical thinking in physics, particularly in magnetism, poses substantial challenges due to the limitations of conventional tools in measuring higher-order cognitive skills. This study aimed to validate an analytical skills test in physics, based on Bloom's Revised Taxonomy, with an emphasis on the dimensions of…
Descriptors: Physics, Science Tests, Science Instruction, Thinking Skills
PaaBen, Benjamin; Dywel, Malwina; Fleckenstein, Melanie; Pinkwart, Niels – International Educational Data Mining Society, 2022
Item response theory (IRT) is a popular method to infer student abilities and item difficulties from observed test responses. However, IRT struggles with two challenges: How to map items to skills if multiple skills are present? And how to infer the ability of new students that have not been part of the training data? Inspired by recent advances…
Descriptors: Item Response Theory, Test Items, Item Analysis, Inferences
Chu, Wei; Pavlik, Philip I., Jr. – International Educational Data Mining Society, 2023
In adaptive learning systems, various models are employed to obtain the optimal learning schedule and review for a specific learner. Models of learning are used to estimate the learner's current recall probability by incorporating features or predictors proposed by psychological theory or empirically relevant to learners' performance. Logistic…
Descriptors: Reaction Time, Accuracy, Models, Predictor Variables
Scott, Terry F.; Schumayer, Dániel – Physical Review Physics Education Research, 2017
The Force Concept Inventory is one of the most popular and most analyzed multiple-choice concept tests used to investigate students' understanding of Newtonian mechanics. The correct answers poll a set of underlying Newtonian concepts and the coherence of these underlying concepts has been found in the data. However, this inventory was constructed…
Descriptors: World Views, Scientific Concepts, Scientific Principles, Multiple Choice Tests
Bulut, Okan; Quo, Qi; Gierl, Mark J. – Large-scale Assessments in Education, 2017
Position effects may occur in both paper--pencil tests and computerized assessments when examinees respond to the same test items located in different positions on the test. To examine position effects in large-scale assessments, previous studies often used multilevel item response models within the generalized linear mixed modeling framework.…
Descriptors: Structural Equation Models, Educational Assessment, Measurement, Test Items
Baghaei, Purya; Dourakhshan, Alireza – International Journal of Language Testing, 2016
The purpose of the present study is to compare the psychometric qualities of canonical single-response multiple-choice items with their double-response counterparts. Thirty, two-response fouroption grammar items for undergraduate students of English were constructed. A second version of the test was constructed by replacing one of the correct…
Descriptors: Language Tests, Multiple Choice Tests, Test Items, Factor Analysis
Bichi, Ado Abdu; Hafiz, Hadiza; Bello, Samira Abdullahi – International Journal of Evaluation and Research in Education, 2016
High-stakes testing is used for the purposes of providing results that have important consequences. Validity is the cornerstone upon which all measurement systems are built. This study applied the Item Response Theory principles to analyse Northwest University Kano Post-UTME Economics test items. The developed fifty (50) economics test items was…
Descriptors: Item Response Theory, Test Items, Difficulty Level, Statistical Analysis
Peer reviewedSerlin, Ronald C.; Kaiser, Henry F. – Educational and Psychological Measurement, 1978
When multiple-choice tests are scored in the usual manner, giving each correct answer one point, information concerning response patterns is lost. A method for utilizing this information is suggested. An example is presented and compared with two conventional methods of scoring. (Author/JKS)
Descriptors: Correlation, Factor Analysis, Item Analysis, Multiple Choice Tests
Peer reviewedPowell, J. C.; Isbister, Alvin G. – Educational and Psychological Measurement, 1974
Descriptors: Factor Analysis, Factor Structure, Information Utilization, Item Analysis
O'Reilly, Robert P.; And Others – 1976
This report summarizes the development and validation of a test of literal comprehension based on a modified cloze technique. This modification of the cloze offers an objective and partially computerized procedure for constructing items in the multiple-choice format. A field test of the multiple-choice cloze (MCC) involved the administration of 36…
Descriptors: Cloze Procedure, Elementary Secondary Education, Factor Analysis, Item Analysis
Ryan, Joseph P.; Hamm, Debra W. – 1976
A procedure is described for increasing the reliability of tests after they have been given and for developing shorter but more reliable tests. Eight tests administered to 200 graduate students studying educational research are analyzed. The analysis considers the original tests, the items loading on the first factor of the test, and the items…
Descriptors: Career Development, Factor Analysis, Factor Structure, Item Analysis
Kingsbury, G. Gage – 1985
A procedure for assessing content-area and total-test dimensionality which uses response function discrepancies (RFD) was studied. Three different versions of the RFD procedure were compared to Bejar's principal axis content-area procedure and Indow and Samejima's exploratory factor analytic technique. The procedures were compared in terms of the…
Descriptors: Achievement Tests, Comparative Analysis, Elementary Education, Estimation (Mathematics)

Direct link
