Publication Date
In 2025 | 20 |
Since 2024 | 68 |
Descriptor
Source
Author
Publication Type
Journal Articles | 63 |
Reports - Research | 61 |
Reports - Descriptive | 3 |
Tests/Questionnaires | 3 |
Information Analyses | 2 |
Reports - Evaluative | 2 |
Dissertations/Theses -… | 1 |
Education Level
Audience
Location
China | 6 |
Indonesia | 2 |
Turkey | 2 |
Europe | 1 |
Iran | 1 |
Nigeria | 1 |
North America | 1 |
Oman | 1 |
Portugal | 1 |
Slovakia | 1 |
South Korea | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
Big Five Inventory | 1 |
International English… | 1 |
Progress in International… | 1 |
Trends in International… | 1 |
What Works Clearinghouse Rating
Daniel M. K. Lam – ELT Journal, 2025
Feedback penetrates many walks of our lives, and its importance in L2 teaching and assessment is well recognised. However, while corrective feedback and writing feedback have been the focus of much L2 research and classroom practice, there seems relatively little attention to feedback on spoken interactional skills. Concomitantly, translating…
Descriptors: Feedback (Response), Peer Evaluation, Oral Language, Interaction
Jesús Pérez; Eladio Dapena; Jose Aguilar – Education and Information Technologies, 2024
In tutoring systems, a pedagogical policy, which decides the next action for the tutor to take, is important because it determines how well students will learn. An effective pedagogical policy must adapt its actions according to the student's features, such as knowledge, error patterns, and emotions. For adapting difficulty, it is common to…
Descriptors: Feedback (Response), Intelligent Tutoring Systems, Reinforcement, Difficulty Level
Yun-Kyung Kim; Li Cai – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2025
This paper introduces an application of cross-classified item response theory (IRT) modeling to an assessment utilizing the embedded standard setting (ESS) method (Lewis & Cook). The cross-classified IRT model is used to treat both item and person effects as random, where the item effects are regressed on the target performance levels (target…
Descriptors: Standard Setting (Scoring), Item Response Theory, Test Items, Difficulty Level
Kala Krishna; Pelin Akyol; Esma Ozer – National Bureau of Economic Research, 2025
Exams are designed to rank students objectively by their abilities, including elements such as time limits, the number and difficulty of questions, and negative marking policies. Using data from a lab-in-field experiment, we develop and estimate a model of student behavior in multiple-choice exams that incorporates the effects of time constraints…
Descriptors: Multiple Choice Tests, Student Behavior, Response Style (Tests), Time
Patrik Havan; Michal Kohút; Peter Halama – International Journal of Testing, 2025
Acquiescence is the tendency of participants to shift their responses to agreement. Lechner et al. (2019) introduced the following mechanisms of acquiescence: social deference and cognitive processing. We added their interaction into a theoretical framework. The sample consists of 557 participants. We found significant medium strong relationship…
Descriptors: Cognitive Processes, Attention, Difficulty Level, Reflection
Aiman Mohammad Freihat; Omar Saleh Bani Yassin – Educational Process: International Journal, 2025
Background/purpose: This study aimed to reveal the accuracy of estimation of multiple-choice test items parameters following the models of the item-response theory in measurement. Materials/methods: The researchers depended on the measurement accuracy indicators, which express the absolute difference between the estimated and actual values of the…
Descriptors: Accuracy, Computation, Multiple Choice Tests, Test Items
Ted Peterson – Journal of Education for Business, 2024
This article delves into student course feedback using publicly available data from the University of North Texas. It examines factors influencing elevated student course evaluation ratings in the Information Technology and Decision Sciences department. The study reveals a positive relationship between higher response rates and better evaluation…
Descriptors: Course Evaluation, Feedback (Response), College Students, Influences
Janea J. Thibodeaux; Pierce M. Taylor; Janelle K. Bacotti; Samuel L. Morris – Journal of Applied Behavior Analysis, 2025
Many researchers have evaluated how characteristics of feedback may influence trainee performance, but relatively little attention has been allocated to directly assessing trainee preference for feedback characteristics and its relation to performance. Thus, the primary purpose of this study was to use a within-subject experimental design to…
Descriptors: Undergraduate Students, Feedback (Response), Difficulty Level, Learning Strategies
Haiyang Yu; Entai Wang; Qi Lang; Jianan Wang – IEEE Transactions on Learning Technologies, 2024
The latest technologies in natural language processing provide creative, knowledge retrieval, and question-answering technologies in the design of intelligent education, which can provide learners with personalized feedback and expert guidance. Entrepreneurship education aims to cultivate and develop the innovative thinking and entrepreneurial…
Descriptors: Entrepreneurship, Comprehension, Questioning Techniques, Information Retrieval
Kuan-Yu Jin; Thomas Eckes – Educational and Psychological Measurement, 2024
Insufficient effort responding (IER) refers to a lack of effort when answering survey or questionnaire items. Such items typically offer more than two ordered response categories, with Likert-type scales as the most prominent example. The underlying assumption is that the successive categories reflect increasing levels of the latent variable…
Descriptors: Item Response Theory, Test Items, Test Wiseness, Surveys
Benjamin M. Torsney; Sarah Rawls; Joseph I. Eisman; Catherine Pressimone Beckowski; Cheryl B. Torsney – Educational and Developmental Psychologist, 2025
Objective: The objective of this study was threefold: (a) to create a rubric for response complexity (RC), defined as an admixture of response length, grammatical diversity, categorisation, and sophistication; (b) to measure behavioural and cognitive engagement through students' written responses on a school-based written activity, and (c) to…
Descriptors: College Students, Learner Engagement, Responses, Difficulty Level
Piotr Jabkowski; Aneta Piekut – Field Methods, 2024
This study analyzes the consequences of item nonresponse to the question about a household's total net income in the European Social Survey (2008-2018). We recognize two mechanisms in avoiding answering the income question: task complexity and question sensitivity, and apply multilevel logistic regressions to predict the probability of refusals or…
Descriptors: Foreign Countries, Family Income, Surveys, Social Characteristics
Neda Kianinezhad; Mohsen Kianinezhad – Language Education & Assessment, 2025
This study presents a comparative analysis of classical reliability measures, including Cronbach's alpha, test-retest, and parallel forms reliability, alongside modern psychometric methods such as the Rasch model and Mokken scaling, to evaluate the reliability of C-tests in language proficiency assessment. Utilizing data from 150 participants…
Descriptors: Psychometrics, Test Reliability, Language Proficiency, Language Tests
Alejandra Ruiz-Segura; Andrew Law; Sion Jennings; Alain Bourgon; Ethan Churchill; Susanne Lajoie – Journal of Computer Assisted Learning, 2024
Background: Flying accuracy is influenced by pilots' affective reactions to task demands. A better understanding of task-related emotions and flying performance is needed to enhance pilot training. Objective: Understand pilot trainees' performance and emotional dynamics (intensity, frequency and variability) based on training phase and difficulty…
Descriptors: Foreign Countries, Flight Training, Aviation Technology, Computer Simulation
Hojung Kim; Changkyung Song; Jiyoung Kim; Hyeyun Jeong; Jisoo Park – Language Testing in Asia, 2024
This study presents a modified version of the Korean Elicited Imitation (EI) test, designed to resemble natural spoken language, and validates its reliability as a measure of proficiency. The study assesses the correlation between average test scores and Test of Proficiency in Korean (TOPIK) levels, examining score distributions among beginner,…
Descriptors: Korean, Test Validity, Test Reliability, Imitation