Publication Date
| In 2026 | 0 |
| Since 2025 | 2 |
| Since 2022 (last 5 years) | 12 |
| Since 2017 (last 10 years) | 22 |
| Since 2007 (last 20 years) | 30 |
Descriptor
Source
| Language Testing | 45 |
Author
| Brown, James Dean | 2 |
| Henning, Grant | 2 |
| Janssen, Gerriet | 2 |
| McCray, Gareth | 2 |
| Trace, Jonathan | 2 |
| Abbott, Marilyn L. | 1 |
| Alderson, J. Charles | 1 |
| Alpizar, David | 1 |
| Aryadoust, Vahid | 1 |
| Bachman, Lyle F. | 1 |
| Baghaei, Purya | 1 |
| More ▼ | |
Publication Type
| Journal Articles | 45 |
| Reports - Research | 39 |
| Tests/Questionnaires | 4 |
| Reports - Descriptive | 3 |
| Reports - Evaluative | 2 |
| Information Analyses | 1 |
| Numerical/Quantitative Data | 1 |
Education Level
| Higher Education | 10 |
| Postsecondary Education | 9 |
| Elementary Education | 2 |
| Secondary Education | 2 |
| Adult Education | 1 |
| Early Childhood Education | 1 |
| High Schools | 1 |
| Kindergarten | 1 |
| Primary Education | 1 |
Audience
Location
| Japan | 4 |
| China | 2 |
| Europe | 2 |
| Iran | 2 |
| Australia | 1 |
| Belgium | 1 |
| China (Guangzhou) | 1 |
| Colombia | 1 |
| Finland | 1 |
| Hong Kong | 1 |
| Malaysia | 1 |
| More ▼ | |
Laws, Policies, & Programs
Assessments and Surveys
| Test of English as a Foreign… | 4 |
| Peabody Picture Vocabulary… | 2 |
| Clinical Evaluation of… | 1 |
| International English… | 1 |
| Test of English for… | 1 |
What Works Clearinghouse Rating
Peer reviewedRaatz, Ulrich – Language Testing, 1985
Argues that classical test theory cannot be used at the item level on "authentic" language tests. However, if the total score is derived by adding the scores of a number of different and independent parts, test reliability can be estimated. Suggests using the Classical Latent Additives model to examine test-part homogeneity. (Author/SED)
Descriptors: Item Analysis, Latent Trait Theory, Models, Second Language Learning
Peer reviewedKim, Mikyung – Language Testing, 2001
Investigates differential item functioning (DIF) across two different broad language groupings, Asian and European, in a speaking test in which the test takers' responses were rated polytomously. Data were collected from 1038 nonnative speakers of English from France, Hong Kong, Japan, Spain, Switzerland, and Thailand who took the SPEAK test in…
Descriptors: English (Second Language), Foreign Countries, Item Analysis, Language Tests
Peer reviewedAlderson, J. Charles; Percsich, Richard; Szabo, Gabor – Language Testing, 2000
Reports on the potential problems in scoring responses to sequencing tests, the development of a computer program to overcome these difficulties, and an exploration of the value of scoring procedures. (Author/VWL)
Descriptors: Computer Software, Foreign Countries, Item Analysis, Language Tests
Song, Min-Young – Language Testing, 2008
This paper concerns the divisibility of comprehension subskills measured in L2 listening and reading tests. Motivated by the administration of the new Web-based English as a Second Language Placement Exam (WB-ESLPE) at UCLA, this study addresses the following research questions: first, to what extent do the WB-ESLPE listening and reading items…
Descriptors: Structural Equation Models, Second Language Learning, Reading Tests, Inferences
Peer reviewedTakala, Sauli; Kaftandjieva, Felianka – Language Testing, 2000
Analyzes gender-uniform differential item functioning (DIF) in a second language vocabulary test with the tools of item response theory to study potential gender impact on the test performance measured by different item composites. Results show that while there are test items with indications of DIF in favor of either females or males, the test as…
Descriptors: English (Second Language), Foreign Countries, Item Analysis, Language Tests
Peer reviewedHenning, Grant – Language Testing, 1988
Violations of item unidimensionality on language tests produced distorted estimates of person ability, and violations of person unidimensionality produced distorted estimates of item difficulty. The Bejar Method was sensitive to such distortions. (Author)
Descriptors: Construct Validity, Content Validity, Difficulty Level, Item Analysis
Abbott, Marilyn L. – Language Testing, 2007
In this article, I describe a practical application of the Roussos and Stout (1996) multidimensional analysis framework for interpreting group performance differences on an ESL reading proficiency test. Although a variety of statistical methods have been developed for flagging test items that function differentially for equal ability examinees…
Descriptors: Test Bias, Test Items, English (Second Language), Second Language Learning
Peer reviewedSpelberg, Henk C. Lutje; de Boer, Paulien; van den Bos, Kees P. – Language Testing, 2000
Compares two language tests with different item types. The tests are the Dutch Reynell test and the BELL test. Both tests were administered to 64 Dutch kindergarten children with an average age of 70.3 months. Regression analyses indicate that item type does not contribute significantly to prediction of item difficulty, but the linguistic…
Descriptors: Comparative Analysis, Dutch, Foreign Countries, Item Analysis
Pae, Tae-Il; Park, Gi-Pyo – Language Testing, 2006
The present study utilized both the IRT-LR (item response theory likelihood ratio) and a series of CFA (confirmatory factor analysis) multi-sample analyses to systematically examine the relationships between DIF (differential item functioning) and DTF (differential test functioning) with a random sample of 15 000 Korean examinees. Specifically,…
Descriptors: Item Response Theory, Factor Analysis, Test Bias, Test Validity
Peer reviewedBrown, James Dean – Language Testing, 1988
The reliability and validity of a cloze procedure used as an English-as-a-second-language (ESL) test in China were improved by applying traditional item analysis and selection techniques. The 'best' test items were chosen on the basis of item facility and discrimination indices, and were administered as a 'tailored cloze.' 29 references listed.…
Descriptors: Adaptive Testing, Cloze Procedure, English (Second Language), Foreign Countries
Peer reviewedReynolds, Trudy; And Others – Language Testing, 1994
Presents a study conducted to provide a comparative analysis of five item analysis indices using both IRT and non-IRT indices to describe the characteristics of flagged items and to investigate the appropriateness of logistic regression as an item analysis technique for further studies. The performance of five item analysis indices was examined.…
Descriptors: College Students, Comparative Analysis, English (Second Language), Item Analysis
Peer reviewedBachman, Lyle F.; And Others – Language Testing, 1996
Discusses the value of content considerations in the design of language tests and the implications of the findings of various investigations of content analysis. The article argues that content analysis can be viewed as the application of a model of test design to a particular measurement instrument, using judgments of trained analysts. (26…
Descriptors: College Students, Content Analysis, English (Second Language), Item Analysis
Peer reviewedHenning, Grant; And Others – Language Testing, 1994
Examines the effectiveness of an automated language proficiency test assembly system at an air force base English Language Center. The study focuses on the equivalence of mean score difficulty, total score variance, and intercorrelation covariance across test norms and finds a high level of test-form equivalence and internal consistency. (nine…
Descriptors: Computer Assisted Testing, English (Second Language), Foreign Nationals, Item Analysis
Peer reviewedChapelle, Carol – Language Testing, 1988
Investigates the relationship between field independence and language measures. Results indicate varying relationships of field independence with cloze, dictation, and multiple-choice language tests. These relationships also differ for native speakers in regular or remedial English classes, and for nonnative speakers. 53 references cited. Cloze…
Descriptors: Cloze Procedure, College Freshmen, Dictation, English (Second Language)
Carr, Nathan T. – Language Testing, 2006
The present study focuses on the task characteristics of reading passages and key sentences in a test of second language reading. Using a new methodological approach to describe variation in test task characteristics and explore how differences in these characteristics might relate to examinee performance, it posed the two following research…
Descriptors: English for Academic Purposes, Sentences, Reading Comprehension, Factor Analysis

Direct link
