Publication Date
| In 2026 | 0 |
| Since 2025 | 2 |
| Since 2022 (last 5 years) | 20 |
| Since 2017 (last 10 years) | 38 |
| Since 2007 (last 20 years) | 65 |
Descriptor
Source
Author
| Paul De Boeck | 3 |
| Amanda Goodwin | 2 |
| Bejar, Isaac I. | 2 |
| Deane, Paul | 2 |
| Katz, Irvin R. | 2 |
| Kostin, Irene | 2 |
| Matthew Naveiras | 2 |
| Papasalouros, Andreas | 2 |
| Saha, Sujan Kumar | 2 |
| Sheehan, Kathleen M. | 2 |
| Solano-Flores, Guillermo | 2 |
| More ▼ | |
Publication Type
Education Level
Audience
| Practitioners | 2 |
| Researchers | 2 |
| Administrators | 1 |
| Teachers | 1 |
Location
| China | 4 |
| Taiwan | 3 |
| Australia | 2 |
| Canada | 2 |
| Germany | 2 |
| Japan | 2 |
| Russia | 2 |
| Alabama | 1 |
| Arizona | 1 |
| Arkansas | 1 |
| California | 1 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Abdullah Faruk Kiliç; Meltem Acar Güvendir; Gül Güler; Tugay Kaçak – Measurement: Interdisciplinary Research and Perspectives, 2025
In this study, the extent to wording effects impact structure and factor loadings, internal consistency and measurement invariance was outlined. The modified form, which includes items that semantically reversed, explains %21.5 more variance than the original form. Also, reversed items' factor loadings are higher. As a result of CFA, indexes…
Descriptors: Test Items, Factor Structure, Test Reliability, Semantics
Brent A. Stevenor; Nadine LeBarron McBride; Charles Anyanwu – Journal of Applied Testing Technology, 2025
Enemy items are two test items that should not be presented to a candidate on the same test. Identifying enemies is essential for personnel assessment, as they weaken the measurement precision and validity of a test. In this research, we examined the effectiveness of lexical and semantic natural language processing techniques for identifying enemy…
Descriptors: Test Items, Natural Language Processing, Occupational Tests, Test Construction
Archana Praveen Kumar; Ashalatha Nayak; Manjula Shenoy K.; Chaitanya; Kaustav Ghosh – International Journal of Artificial Intelligence in Education, 2024
Multiple Choice Questions (MCQs) are a popular assessment method because they enable automated evaluation, flexible administration and use with huge groups. Despite these benefits, the manual construction of MCQs is challenging, time-consuming and error-prone. This is because each MCQ is comprised of a question called the "stem", a…
Descriptors: Multiple Choice Tests, Test Construction, Test Items, Semantics
Qiao, Chen; Hu, Xiao – IEEE Transactions on Learning Technologies, 2023
Free text answers to short questions can reflect students' mastery of concepts and their relationships relevant to learning objectives. However, automating the assessment of free text answers has been challenging due to the complexity of natural language. Existing studies often predict the scores of free text answers in a "black box"…
Descriptors: Computer Assisted Testing, Automation, Test Items, Semantics
Frank Goldhammer; Ulf Kroehne; Carolin Hahnel; Johannes Naumann; Paul De Boeck – Journal of Educational Measurement, 2024
The efficiency of cognitive component skills is typically assessed with speeded performance tests. Interpreting only effective ability or effective speed as efficiency may be challenging because of the within-person dependency between both variables (speed-ability tradeoff, SAT). The present study measures efficiency as effective ability…
Descriptors: Timed Tests, Efficiency, Scores, Test Interpretation
Schaper, Marie Luisa; Kuhlmann, Beatrice G.; Bayen, Ute J. – Journal of Experimental Psychology: Learning, Memory, and Cognition, 2023
Item memory and source memory are different aspects of episodic remembering. To investigate metamemory differences between them, the authors assessed systematic differences between predictions of item memory via Judgments of Learning (JOLs) and source memory via Judgments of Source (JOSs). Schema-based expectations affect JOLs and JOSs…
Descriptors: Memory, Metacognition, Schemata (Cognition), Prediction
Jordan M. Wheeler; Allan S. Cohen; Shiyu Wang – Journal of Educational and Behavioral Statistics, 2024
Topic models are mathematical and statistical models used to analyze textual data. The objective of topic models is to gain information about the latent semantic space of a set of related textual data. The semantic space of a set of textual data contains the relationship between documents and words and how they are used. Topic models are becoming…
Descriptors: Semantics, Educational Assessment, Evaluators, Reliability
Becker, Kirk A.; Kao, Shu-chuan – Journal of Applied Testing Technology, 2022
Natural Language Processing (NLP) offers methods for understanding and quantifying the similarity between written documents. Within the testing industry these methods have been used for automatic item generation, automated scoring of text and speech, modeling item characteristics, automatic question answering, machine translation, and automated…
Descriptors: Item Banks, Natural Language Processing, Computer Assisted Testing, Scoring
Sun-Joo Cho; Amanda Goodwin; Matthew Naveiras; Paul De Boeck – Grantee Submission, 2024
Explanatory item response models (EIRMs) have been applied to investigate the effects of person covariates, item covariates, and their interactions in the fields of reading education and psycholinguistics. In practice, it is often assumed that the relationships between the covariates and the logit transformation of item response probability are…
Descriptors: Item Response Theory, Test Items, Models, Maximum Likelihood Statistics
Sun-Joo Cho; Amanda Goodwin; Matthew Naveiras; Paul De Boeck – Journal of Educational Measurement, 2024
Explanatory item response models (EIRMs) have been applied to investigate the effects of person covariates, item covariates, and their interactions in the fields of reading education and psycholinguistics. In practice, it is often assumed that the relationships between the covariates and the logit transformation of item response probability are…
Descriptors: Item Response Theory, Test Items, Models, Maximum Likelihood Statistics
C. H., Dhawaleswar Rao; Saha, Sujan Kumar – IEEE Transactions on Learning Technologies, 2023
Multiple-choice question (MCQ) plays a significant role in educational assessment. Automatic MCQ generation has been an active research area for years, and many systems have been developed for MCQ generation. Still, we could not find any system that generates accurate MCQs from school-level textbook contents that are useful in real examinations.…
Descriptors: Multiple Choice Tests, Computer Assisted Testing, Automation, Test Items
Peter Organisciak; Selcuk Acar; Denis Dumas; Kelly Berthiaume – Grantee Submission, 2023
Automated scoring for divergent thinking (DT) seeks to overcome a key obstacle to creativity measurement: the effort, cost, and reliability of scoring open-ended tests. For a common test of DT, the Alternate Uses Task (AUT), the primary automated approach casts the problem as a semantic distance between a prompt and the resulting idea in a text…
Descriptors: Automation, Computer Assisted Testing, Scoring, Creative Thinking
Reima Al-Jarf – Online Submission, 2024
Expressions of impossibility refer to events that can never or rarely happen, tasks that are difficult or impossible to perform, people or things that are of no use and things that are impossible to find. This study explores the similarities and differences between English and Arabic expressions of impossibility, and the difficulties that…
Descriptors: English (Second Language), Second Language Learning, Arabic, Translation
Brainerd, C. J.; Bialer, D. M.; Chang, M.; Upadhyay, P. – Journal of Experimental Psychology: Learning, Memory, and Cognition, 2022
In recognition memory, anything that is objectively new is necessarily not-old, and anything that is objectively old is necessarily not-new. Therefore, judging whether a test item is new is logically equivalent to judging whether it is old, and conversely. Nevertheless, a series of 10 experiments showed that old? and new? judgments did not produce…
Descriptors: Memory, Recognition (Psychology), Novelty (Stimulus Dimension), Evaluative Thinking
Kanj, Rama; El-Hassan, Karma – International Journal of Multilingualism, 2023
Vocabulary tests administered on multilingual populations should take into account the unique linguistic and cultural makeup of the population by adopting test development methods that allow responses in several languages. Our aims were to develop a picture-naming test for multilingual Lebanese school-age children (L1: Lebanese, L2: French and/or…
Descriptors: Vocabulary Development, Language Tests, Expressive Language, Multilingualism

Peer reviewed
Direct link
