One important step in selecting texts for comprehension assessment is to evaluate the suitability of the vocabulary content for the learners' level of proficiency in the language, since it is unreasonable to expect them to understand a text containing a substantial number of unknown lexical items. Traditionally, this step is assisted by applying a standard readability formula, such as the Flesch Reading Ease score or the Flesch–Kincaid Grade Level score (both available in Microsoft Word), incorporating word frequency as a core component. Another approach is to submit the text to the VocabProfile section of the Compleat Lexical Tutor (www.lextutor.ca), which offers both color coding and frequency statistics to distinguish common words from those that occur less frequently. It should be noted that both these approaches are word based, so they may underestimate the lexical difficulty of texts containing idiomatic or colloquial expressions.
Vocabulary assessment for comprehension purposes is embedded, in the sense that it engages with a larger construct than just vocabulary knowledge or ability. In practical terms, this means that the vocabulary test items form a subset of items within the test. In addition, the focus of the items changes from simply eliciting evidence of the ability to recognize or recall word meanings to contextual understanding, through reading items like these:
The word “inherent” in line 17 means. ..
Find a phrase in paragraph 3 that means the same as “analyzing.”
Items may also assess lexical‐inferencing ability by targeting vocabulary items that the test takers are unlikely to know, but whose meaning can reasonably be inferred by clues available in the surrounding text.
By its nature, reading assessment offers more scope for items that focus on individual lexical units in the text than listening assessment, because a written text remains available for review in a way that a spoken text does not. In fact, it may be counterproductive to encourage learners to concentrate on vocabulary in a listening task, as Chang and Read (2006) found in their investigation of various forms of support for EFL learners taking a listening test. Pre‐teaching of key lexical items proved to be the least effective form of support, apparently because it drew the test takers' attention away from the propositional content of the text. On the other hand, if spoken or written texts represent a particular discipline, register, or genre, comprehension test items will provide at least indirect evidence of the ability to handle vocabulary appropriate for that type of text (see Read, 2007, for further discussion of this point), even if none of the test items focus explicitly on vocabulary.
Use
Use refers to the ability of learners to draw on their vocabulary resources in undertaking speaking or writing tasks, like giving a talk, participating in a conversation or discussion, composing a letter, writing an essay, or compiling a report. This represents a more genuine sense of production than a recall task such as supplying a content word to complete a gap in a sentence. One characteristic of vocabulary use tasks which distinguishes them from the other three approaches outlined above is that the task designer cannot normally target particular lexical items by requiring the learners to incorporate specific words into what they produce. Thus, the choice of words can only be influenced indirectly by the choice of task and topic, or by the selection of appropriate input material: source texts, pictures, diagrams, and so on.
As with comprehension tasks, use tasks can be assessed purely as measures of the learners' vocabulary ability or as measures of a larger speaking or writing construct in which vocabulary is embedded as one component. Vocabulary researchers have devised a variety of statistics to evaluate the lexical characteristics of texts: How many different words are used? What percentage of the words are low‐frequency items? What percentage are content words? Until now, the statistics have not been very practical tools for assessment purposes but recent advances in automated writing assessment (Carr, 2014), in which lexical measures play a prominent role, mean that automated ratings complement human judgments in the assessment of writing in the Internet‐based Test of English as a Foreign Language (TOEFL), and they completely replace human raters in the Pearson Test of English (Academic).
For now the more common practice in speaking and writing tasks is for the teacher, or the rater in the case of a more formal test, to assess the learners' use of vocabulary by means of a rating scale. For example, in the speaking module of the International English Language Testing System (IELTS), lexical resource is one of four criteria that the examiners apply to each candidate's performance, along with fluency and coherence, grammatical range and accuracy, and pronunciation. Highly proficient candidates are expected to use a wide range of vocabulary accurately and idiomatically, whereas those with more limited speaking proficiency are restricted to talking about familiar topics and lack the ability to paraphrase what they want to say. Thus, such assessments are based on raters' perceptions of general lexical features of the test takers' task performance, rather than on any individual vocabulary items.
Conclusion
Given the central role of vocabulary in language acquisition and use, a whole range of procedures is needed to assess learners' developing vocabulary ability in a second language. Much depends on the purpose of the assessment. For beginning and intermediate learners, who are building their core knowledge of high‐frequency words, recognition and recall formats have an important role in monitoring their lexical development. Such formats are also useful in placement and diagnostic tests. However, as learners' competence in the second language advances, we need to assess their ability to draw on their vocabulary resources effectively for functional communication by means of comprehension and use tasks.
SEE ALSO: Assessment of Reading; Assessment of Writing; Corpus Linguistics in Language Teaching; Formulaic Language and Collocation; Teaching Vocabulary; Vocabulary and Language for Specific Purposes
References
1 Ackermann, K., & Chen, Y. (2013). Developing the academic collocations list (ACL): A corpus‐driven and expert‐judged approach. Journal of English for Academic Purposes, 12(4), 235–47.
2 Brezina, V., & Gablasova, D. (2015). Is there a core general vocabulary? Introducing the new general service list. Applied Linguistics, 36(1), 1–22.
3 Browne, C., Culligan, B., & Phillips, J. (2013). A new general service list. Retrieved April 2, 2019 from www.newgeneralservicelist.org/
4 Carr, N. T. (2014). Computer‐automated scoring of written responses. In A. J. Kunnan (Ed.), The companion to language assessment (Vol. 2, chap. 64). Chichester, England: Wiley‐Blackwell.
5 Chang, A. C., & Read, J. (2006). The effects of listening support on the listening performance of EFL learners. TESOL Quarterly, 40(2), 375–97.
6 Coxhead, A. (2000). A new academic word list. TESOL Quarterly, 34(2), 213–38.
7 Dang, T. N. Y., Coxhead, A., & Webb, S. (2017). The academic spoken word list. Language Learning, 67(4), 959–97.
8 Elgort, I. (2011). Deliberate learning and vocabulary acquisition in a second language. Language Learning, 61(2), 367–413.
9 Gardner, D., & Davies, M. (2014). A new academic vocabulary list. Applied Linguistics, 35(3), 305–27.
10 Nation, I. S. P. (2013). Learning vocabulary in another language