Reicher (1969) and Wheeler (1970) controlled for response bias by asking participants which of two letters had been briefly presented (and masked) in a particular position. For example, given the string lake, probing whether k or t had appeared in the third position would not produce a word bias, because either letter completes a word. The publication of these experiments stimulated a generation of research on the “word superiority effect,” eventually leading to a modified conclusion: Letters within nonword pronounceable strings (pseudowords) are also perceived better than random strings of letters. Letters in real words are perceived a little better than letters in these pseudowords, but the largest difference seems to concern the internal structure of the letter string, its word‐like orthography and phonology.
McClelland and Rumelhart (1981) explained both the word superiority effect and the pseudoword superiority effect in a new approach, a model that connected three hierarchical levels – words, letters, and letter segments – with bi‐directional activation between adjacent levels of the hierarchy. Activation spreading from letters up to words accumulates recognition evidence for specific words; and activation from a word down to the letter level accumulates evidence for the letters in that word. Thus, letters are perceived better in pseudowords than letter strings because they receive feedback from words that contain these letters (e.g., the k in loke receives feedback from lake and like). Similarly, bi‐directional activation causes k to be better perceived in lake than loke, producing word superiority effect.
This approach became a model for how to conceptualize “interaction” in a precise way. The explicit representation of letters and words in a lexical memory system later gave way to Parallel Distributed Processing (PDP) models that learned connections rather than having them built in (Plaut et al., 1996; Seidenberg & McClelland, 1989; Seidenberg et al., this volume). However, the principles of the original interactive model with “localized” lexical representations were retained in other models of alphabetic reading (e.g., Grainger & Jacobs, 1996). Many computational models have been developed since these earlier models, which were restricted by small lexicons and limited generality across word reading tasks (Norris, 2013). These problems, and the focus on alphabetic writing, continue to challenge the generality of reading models.
The lexicon and how to get there from an orthographic string.
The distinction between computing and retrieving word pronunciations has had an enduring influence on models of reading. Early expressions of dual route ideas (Baron & Strawson, 1976; Forster & Chambers, 1973) became formalized by Coltheart et al. (2001) in the Dual Route Cascaded (DRC) model: A reader can arrive at a word’s pronunciation in two ways: 1) Decoding its letters to phonemes and producing the aggregated results – the computed route (also called sublexical, assembled, indirect). 2) Retrieving the pronunciation stored with its orthographic word‐form – the retrieved route (also called lexical, addressed, direct).
For a skilled reader, the difference between the two routes escapes notice because reading experience has established familiar lexical representations for many words. Thus, with appropriate experience, a reader may pronounce choir as easily as chore, unaware that the first resulted from the retrieval of a stored pronunciation associated with its spelling, while the second might have resulted from either route depending on familiarity with the word.
Both the DRC and PDP models can simulate word reading performance. For PDP models, the structure of mental representations emerges from many cycles of pattern association and error‐reduction learning. The DRC model, in the tradition of classic models with fixed assumptions, predicts experimental data based on a fixed architecture. Coltheart et al. (2001) showed that dual route models provide many specific, correct predictions of experimental results. The fundamental difference between the two models is between a model that learns – without necessarily showing either the time course or the pattern of learning outcomes of an actual learner – and a model that has already learned and is now ready to read any word or letter string one can throw at it. Narrowing the gap between these models are approaches that add a learning component to the DRC model (Pritchard et al., 2016) and combine elements of connectionist and DRC modeling (Perry et al., 2007; see Seidenberg et al., this volume for discussion).
Phonology in skilled word identification.
Concerning readers’ self‐reports, Huey wrote, “Of nearly thirty adults who were thus tested, the large majority found inner speech in some form to be a part of their ordinary reading. Purely visual reading was not established by any of the readers,.…” (1908, p. 119). This conclusion about phonology during silent reading continues to seem correct (see Brysbaert, this volume).
The issue in word identification is more specific: whether the phonology of a word is “prelexical” – the phonemes activated by letters and letter strings lead to word identification – or “postlexical” – word phonology follows after access to the orthographic form of the word. Opinion generally favored a direct‐to‐meaning identification procedure with no prelexical phonology in skilled reading, rationalized partly by questionable assumptions about the consequences of English spellings: Because English spelling‐to‐pronunciation mappings have inconsistencies, readers learn to read English without using these unreliable mappings.
However, various experimental approaches provided evidence to the contrary. One was to expose a word briefly (35–45 ms) followed by a backward mask consisting of letter strings. When the letter mask reinstated the word’s phonemes, identification of the word improved, even when the letters were changed (choir – #### – kwire) (Perfetti et al., 1988). This effect implies that, prior to the word’s identification, some of its phonology had been activated. Lukatela and Turvey (1994a, b), using a similar logic with primed lexical decision, found that homophone primes (e.g., towed – toad) produced strong facilitation relative to spelling controls. These conclusions were supported by a meta‐analysis by Rastle and Brysbaert (2006).
The most well‐known evidence came from the semantic category judgment experiments of van Orden (1987). Presented with the category “flowers,” readers sometimes made category mistakes on the word rows, suggesting that the word’s phonology was activated automatically, creating confusion with rose. Jared and Seidenberg (1991) found this effect was limited to low frequency words when only shallow meaning (animate/inanimate) decisions were required. For a familiar word, some general meaning features may be accessible prior to full phonology. More generally, both phonological and semantic activations are triggered by a familiar word form in an interdependent way. The rapid activation of a word’s phonology can stabilize the word’s identity including its meaning features (van Orden et al., 1990).
Table 1.1 summarizes the properties and functions of the phonology that, on our account, are part of word identification. In alphabetic reading, this involves automatic, recurring interactions between letter strings and phoneme strings, including the whole word level. These orthographic‐phonological interactions occur in the most rapid swirls of the fast current of skilled reading, resulting in a stable word identity that remains accessible during the reading of the sentence that contains it.
Beyond the mere activation of lexical phonology is its content. The speed of silent reading could suggest that, rather than a fully specified pronunciation, a phonological skeleton of (more reliable) consonants is quickly activated, followed by (less reliable) vowels (Berent & Perfetti, 1995). Other research implicates a fuller, multilevel phonology including stress patterns (Ashby & Clifton, 2005). Some uncertainty remains concerning the phonological content and the time course of segmental (consonants and vowels) and supra‐segmental (lexical stress) phonology. However, a rapidly activated phonological component of word identification has been confirmed in research on sentences as well as isolated words across multiple methods, including eye‐tracking, ERP,