You are on page 1of 37

1

UALL 2004
PSYCHOLINGUISTICS

LECTURE 7

Topic 4: Word Selection and Recognition


2

Semantics Phonology

Selecting Words

Syntax
3

Stepping-stone Waterfall /
Model Cascade Model

Selecting Words

Interactive/Spreading
Activation Model
Stepping-stone Model 4
Stepping-stone Model 5

 Two major stones: (1) meaning and word class, (2) sounds

 First, a person picks a meaning, a word class, and then the sound
structure.

 Example: Looking at a smallish wild animal and later identifying it as a


BEAVER.

 However, there would be a signpost saying whereabouts to go in the


next component: a phonological ‘area-code’.
Stepping-stone Model 6

 ‘BEAVER’ phonological area-code: a two-syllable word beginning with


/b/ and ending in /ər/.

 Yet, the semantic component activates other small wild animals


belonging to the same category: OTTER, BADGER, RABBIT

 Phonologically, words like ‘beaker’, ‘badger’, and ‘bearer’ are activated


too.

 This explains the error of assuming a BEAVER as an OTTER or calling


it a BEARER.
Stepping-stone Model 7
Waterfall/Cascade Model 8
Waterfall/Cascade Model 9

(McClelland, 1979)
 There was a need for a model to indicate that a person still thinks
about the meaning as he/she selects a sound – The waterfall model
shows this.

 In this model, all the information activated at the first stage is still
available at the next stage.

 So, after selecting a word's meaning from the activated group of


words, all of them remain available as the person deals with the
sounds.

 So, word selection is not just a case of following one word through
from beginning to end.
Waterfall/Cascade Model 10
Waterfall/Cascade Model 11

(McClelland, 1979)
 Example:

 A: Think up the names of some woodland animals.


 B: Rabbit, squirrel…I can’t think of anymore.
 A: Hmm….beginning with ‘b’.
 B: Oh….badger, beaver…

 The dialogue shows information can flow both ways: particular sounds
can enable a speaker to activate meanings, just as meanings activate
sounds.
Interactive/Spreading 12

Activation Model
 In speech production, the current is normally initiated in the semantic
component, where a semantic field will be aroused and then narrowed
down.

 The current flows to the phonological ‘area-code’ before the final choice
is made, where many words will be triggered, and those activated will
feed back into semantics (information flow backwards and forwards).

 All the links between activated sections will metaphorically be lit up, with
electric current rushing backwards and forwards, exciting more and more
related words.
Interactive/Spreading 13

Activation Model
Interactive/Spreading 14

Activation Model
 The related words are stimulated strongly while the unrelated ones fade.

 Thus, the activated word groups gradually narrow to more likely


candidates.

 Phonology activation follows the same pattern of activating any that fits
more than one animal and is later matched with semantically activated
meaning to narrow down the choice.

 If a person doesn’t pay much attention, the wrong choice can be made.

 Different words require different activation level.


Word Recognition 15

Basic Problems:

1. It isn’t easy to hear each


phoneme in normal speech.

2. Neighboring sounds alter sounds.

3. Sound segments cannot be


separated – they melt into each
other.
Spoken Word Recognition 16

Process

 (1)Initial contact, (2)lexical decision (selection), and


(3)word recognition are the three stages involved in the
spoken word recognition process.

 Initialcontact: The listener takes the speech as input and


generates the representations that contact the internally
stored form-based representations associated with each
lexical entry.
Spoken Word Recognition 17

Process
 Lexical decision (selection): Accumulating sensory input
continues to map onto this subset until the intended lexical entry
is eventually selected.

 Word recognition – the endpoint of the selection phase when a


listener has determined which lexical entry was heard.
Word Recognition 18

Logogen Model TRACE Model COHORT Model

First Generation Model Second Generation Models

WORD RECOGNITION

Third Generation Model

Simple Recurrent Network Model


Logogen Model 19

John Morton (1969)


 A bottom-up driven system that takes spoken or visual input and uses it
to activate previously stored word form representations.

 A set of processing units that would receive input from spoken or written
modalities and fire when their activated inputs exceeded some criterion
level (threshold).

 A logogen represents each word in a person’s vocabulary, so words are


recognized when the activation levels of their corresponding logogen
exceed the threshold.
Logogen Model 20

John Morton (1969)


Second Generation Models 21

(Models of Speech Recognition – Interactive Models)

 The Cohort Model and TRACE Model are concerned with


auditory word recognition.

 For these models, the result is a meaningful utterance


rather than a meaningless syllable.

 These models aim to describe the interaction between the


processes of phoneme recognition and word recognition.
Cohort Model
(Marslen-Wilson, 1987; Marslen-Wilson & Welsh, 1978;
22
Marslen-Wilson & Tyler, 1980, 1981)

This model was


designed to account
only for auditory word
recognition.
Cohort Model 23
(Marslen-Wilson, 1987; Marslen-Wilson & Welsh, 1978;
Marslen-Wilson & Tyler, 1980, 1981)

Involves three stages: STAGE 1: The acoustic-phonetic


information at the beginning of a target
word activated all words that resembles
Stage 1: Activation (or Contact) it, making up a “word-initial cohort”
 John was trying to get some bottles
down from the top shelf. To reach
Stage 2: Selection them he headed to sta… stab, stack,
stagger, stagnate…
STAGE 2: All other sources of
Stage 3: Integration information (frequency, context) may
influence the selection of the target
word from the cohort
STAGE 3: Final word chosen when
only a single candidate remains in the
cohort.
Cohort Model 24
(Marslen-Wilson, 1987; Marslen-Wilson & Welsh, 1978;
Marslen-Wilson & Tyler, 1980, 1981)

 COHORT views contact as being influenced only by bottom-up auditory


information and not by contextual information.

 So, activation in COHORT is an autonomous process – it is affected by


auditory stimulation but not other potentially relevant cognitive processes.

❑ Thus, stored representations of words that do not fit into the evolving
context are activated anyway as long as they match the acoustic
properties of the word stimulus.
Cohort Model 25
(Marslen-Wilson, 1987; Marslen-Wilson & Welsh, 1978;
Marslen-Wilson & Tyler, 1980, 1981)

 The selection depends on the bottom-up stimulus because bottom-up


information activates word candidates, but it also depends on context.

❑ Words that fit better into the context will have an advantage over words
that do not fit, especially in cases where the bottom-up input is
ambiguous between two or more stored word candidates.

 During integration, properties of the selected word – its grammatical class


and meaning – are evaluated to how well they fit with the preceding
context.
Cohort Model 26
(Marslen-Wilson, 1987; Marslen-Wilson & Welsh, 1978;
Marslen-Wilson & Tyler, 1980, 1981)

 COHORT is called as such because the process of lexical access starts


with a contact phase in which all words that match the perceived acoustic
profile are activated.

 So, within about 100-150ms of the onset of a word, a whole group of


matching candidate word forms become more available or accessible
than usual.

 This group of activated word forms is called a cohort.


Cohort Model 27
(Marslen-Wilson, 1987; Marslen-Wilson & Welsh, 1978;
Marslen-Wilson & Tyler, 1980, 1981)

 One advantage of the model is that it makes very specific predictions


about when, exactly, a word can be recognised and its meaning
accessed.

 Word recognition depends on reducing the set of activated words to the


one that matches the acoustic input. The point where the COHORT is
reduced to a sole survivor is called the ‘recognition point’. (Marslen-
Wilson, 1987).

 The model allows for minor adjustments to the recognition point based on
semantic or syntactic requirements imposed by the context, which allows
words that are highly predictable in context to be recognized faster than
the less predictable ones.
Cohort Model 28
(Marslen-Wilson, 1987; Marslen-Wilson & Welsh, 1978;
Marslen-Wilson & Tyler, 1980, 1981)

 Word recognition in this model is contingent on two factors:

 (1) There has to be positive evidence for the presence of the word (e.g.:
the input ‘tres’ provides clues that the word ‘trespass’ is the matching
word target).

 (2) The input has to rule out the presence of other words (e.g.: the onset
‘tr’ rules out the possibility that the matching word target is ‘tap’, ‘top’
‘table’ or any other word that does not begin with ‘tr’)
TRACE Model 29
(McClelland & Rumelhart, 1981)

(a) indicates the basic organization of


processing units and information flow,
while (b) shows a detailed view of the
way the processing units are
connected to one another.
TRACE Model 30
(McClelland & Rumelhart, 1981)

 A highly interactive model in which activation of processing units is


determined solely by stimulation provided by the input.

 The activation of one processing unit in a bottom-up system does not


directly affect the activation of other processing units at the same system
level.

 Activation at higher levels of a bottom-up processing system does not


affect activation at lower levels.

 Phonemes affect the activation of word units, but word units do not affect
activation in the units that represent phonemes.
TRACE Model 31
(McClelland & Rumelhart, 1981)

 Interactive processing systems have connections between processing


units that allow units within the same level to affect one another and that
allow processing units at higher system levels to affect units at lower
levels.

 The model assumes that activation is ‘cascaded’, contrasting threshold


activation.

 In cascaded activation, units receiving input begin to send output as soon


as any activation at all comes in from other units.

 Visual features start to send activation forward as soon as they begin to


be identified, so letter-level processing units start to become active soon
after feature-level processing units start to become active.
TRACE Model 32
(McClelland & Rumelhart, 1981)

 That means letter representations start to become activated as soon as


any visual feature has been identified, and you do not need to perceive
all of the features of a letter before you start to activate letter-level
processing units.

 The model offers a good explanation of the “word superiority” effect that
indicates that we have an easier time recognising and processing letters
and phonemes when they appear in the context of a word than when they
occur by themselves or in the context of a string of letters that does not
make up a real word.
Cohort and TRACE Models 33
Distributed Feature Models 34

 The parallel distributed processing idea continued to grow and develop


with the invention of newer and more advanced mathematical models of
lexical access.

 Jeff Elman’s Simple Recurrent Network (SRN) model assumed that


words were represented as a pattern of neural activity across multi-
layered network.
Simple Recurrent Network Model
Jeff Elman (2004) 35
Simple Recurrent Network Model 36
Jeff Elman (2004)
 This model took the TRACE model and added a set of context units.

 The context units store a copy of the activations in the hidden units
between processing cycles.

 The network would respond not just to the current state of the input units
but also to recent events, as reflected in the activity of the context units.

 The explicit task that the network performed was to predict the upcoming
word in an utterance.
Simple Recurrent Network Model 37
Jeff Elman (2004)
 In this system, word identities can be represented as an activation
pattern among the hidden units.

 The patterns split neatly into two classes, corresponding to nouns and
verbs; within each class, the word representations are subdivided further
into subclasses, with similar representations assigned to words that we
would judge as being close in meaning.

You might also like