Scovel كتاب PDF
Scovel كتاب PDF
Scovel كتاب PDF
First words
After crying, cooing, and babbling, we come to the culmination of
a child's early language development-the first word. A child
crosses this linguistic Rubicon at about one year old, although
there is a wide range of latitude as to when the first word emerges
and as to what constitutes a 'word'. For one thing, it seems that
children often use idiomorphs, words they invent when they first
catch on to the magical notion that certain sounds have a unique
reference. For example, one psycholinguist recorded that when
his daughter was about one year old, she came up with 'ka ka' as
the word for 'milk', But just as frequently, youngsters begin to
learn the vocabulary of their mother tongue straight away. A survey
of the words children first learn to say shows that they tend to
be those which refer to prominent, everyday objects, and usually
things that can be manipulated by the child. Thus, 'mama' and
dada' (of course), and 'doggie', 'kitty', but also 'milk', 'cookie,
and 'sock', Even at this most rudimentary stage of vocabulary
development, we can see evidence for what Piaget calls egocentric
speech. Children, quite naturally, want to talk about what surrounds
them; at life's beginnings, they are the center of their universe.
If the child cannot manipulate the object during this early
period of physical development, it does not appear to be worth
naming. Parents spend a lot of time putting diapers on and taking
them off their one-year-olds, but because babies themselves (quite
fortunately!) don't handle them, 'diapers' or 'nappies' do not
become part of a child's early linguistic repertoire.
Parents fuss a great deal over their child's first word; this, and
the first step, rank as singular benchmarks of maturation. The
first cry, the first coo, or the first babble is often ignored or unrecognized,
but the first substantive evidence of vocabulary acquisition,
even if indistinguishable from a controlled burp to outsiders,
is often duly recorded and dated by proud parents. Just as the first
steps are symbolic of the evolution of man from ape-like animal
to biped, the first few words, whether idiomorphs or words from
the parent's native language, demonstrate to the mother and
father that their child has successfully made the transition from
an iconic creature to a symbolic human being.
The Miracle Worker, the compelling drama about the early life
of Helen Keller, saves this marvellous moment for its powerful
conclusion. Annie Sullivan, the teacher hired to transform the
blind and deaf, asocial and non-communicative young Helen, has
been laboring throughout the play to get Helen to communicate
by finger spelling, but now, with Annie's contract almost up, all
seems hopeless. Helen remains entrapped in an iconic world without
speech or language. But as they stand in the well-house, next
to the water pump, where Annie has led Helen for her daily chore
of filling the pitcher for dinner, the water spills accidentally on
Helen's hands and the miracle unfolds. Helen seizes Annie's hand
and finger-spells what Annie has written so many times on
Helen's hand, apparently without success. W-A-T-E-R. From this
moment on, words cascade onto Helen's fingers like the water
which is accidentally spilt at the well; and from this moment
comes an explosion of linguistic learning, so that Helen is eventually
able to write about the experience in her own words.
That living word awakened my soul, gave it light, hope, joy,
set it free... I left the well-house eager to learn. Everything
had a name, and each name gave birth to a new thought.ioi
(from Helen Keller. 1903. The Story of My Life. Doubleday,
page 44)
More remarkable than the drama, and the actual biographical
anecdote it depicts, is that most of us have experienced a
similar moment when, at about the age of one, we too suddenly
recognized the mystic harmony, linking sense to sound and
sight', and entered the sentient and symbolic world of human
communication. Once the first few words are acquired, there is
an exponential growth in vocabulary development, which only
begins to taper at about the age of six, when, by some estimates,
the average child has a recognition vocabulary of about I4,000
words. It is no wonder then that parents are excited by their
child's first word: it represents a step into symbolic communication,
and it signifies the start of the rapid vocabulary growth
with which thoughts, feelings, and perceptions, as well as other
areas of linguistic development, are framed.
Childish creativity
There is another way in which child language acquisition is relatively
independent from environmental influences, despite the distinct
control that the latter exercise on the course of our first
language development. Obviously, a child's linguistic surroundings
determine its mother tongue: children raised in Shandong,
China, grow up speaking Mandarin; children raised in
Bedfordshire, England, grow up as native speakers of English;
and children, like your author, who grow up in Shandong but are
reared by native speakers of English, usually acquire bilingual
proficiency in both of these tongues. But despite the obvious
impact the environment has on the choice and general direction of
mother-tongue learning, children are prone to come up with all
kinds of words and expressions which they have never heard in
their mono- or bilingual environments. Children are creative
wordsmiths, as evidenced in the following exchange between a
friend and her two-year-old.
Conceptualization
Where does the very beginning of any spoken utterance come
from? What sparks speech? These are difficult questions to
answer, partly because we still don't know enough about how
language is produced, but partly because they deal with mental
abstractions so vague that they elude empirical investigation. The
American psycholinguist David McNeill, however, has gone on
record with an interesting mentalistic account of how speech is
first conceptualized in the human mind. His theory is that primitive
linguistic concepts are formed as two concurrent and parallel
modes of thought. These are syntactic thinking, which spawns the
sequence of words which we typically think of when we talk
about how language is initiated, and imagistic thinking, which
creates a more holistic and visual mode of communication. The
former is segmented and linear and creates the strings of syllables,
words, phrases, and sentences that together make up speech.
The latter is global and synthetic and tends to develop the
gestures which we naturally use to punctuate and illustrate our
conversations.
McNeill's claim, that syntactic thought and imagistic thought
collaborate to conceptualize conversation, is quite convincingly
demonstrated by the way in which speech utterances and ordinary
gestures seem to be tied and timed together in any conversation.
Consider the following very simple example. Two people are
holding a short discussion over the whereabouts of a lost object.
Visualize in your mind how they gesture as they interact in the following
two dialogues. You might even try reading these aloud,
acting out Person B's role by pointing at the appropriate moment.
First dialogue
Person A: Where's my briefcase?
Person B: There's your briefcase!
Person B points to the briefcase the same moment he says
There's.
Second dialogue
Person A: Where's my coat and briefcase?
Person B: There's your briefcase!
Person B points to the briefcase the same moment he says
briefcase.
What are the very first things that are going through Person B's
mind when she is responding to Person A's questions in these two
dialogues? Of course we cannot be too mentalistic and pretend
we know what B is thinking. After all, we are often unsure of
what we are thinking ourselves when we think about what we
think, if we think about thinking at all. This is the problem with
mentalism. But McNeill offers some plausible evidence for this
bimodal view of how speech is produced. It seems likely that after
B hears A's query in the first example, her syntactic thought might
generate something that begins with the demonstrative, 'there
while, simultaneously, her imagistic thought might be of someone
pointing toward an object, in this case, a briefcase. Evidence that
these two modes are operating concurrently at the conceptualization
stage is found in the simultaneous timing of the pointing gestures
with the stressed words in each of these two scenes. In the
first dialogue, B points to the briefcase (manifesting the imagistic
part of her attempt to communicate) just as she stresses the word
there' in her speech (illustrating the syntactic component of her
communicative intent). Again, in the second dialogue, we see the
synchrony of image and speech; at the end of the phrase B points
to the briefcase just as she stresses the word in her articulation. If
you read this last example out loud, you will also note a slight
change in B's intonation-the voice trails off a bit as if to say
‘There's your briefcase…’ Were B suddenly to spot the coat, she
could continue with 'and there's your coat', with a more decisive,
falling intonation on 'coat' and, of course, another pointing gesture
to show A where his coat was located.
Appealing as McNeill's hypothesis might appear, and convincing
as these examples might be, it is difficult to use his model to
explain this first stage of production. For one thing, his attempts
to describe how imagistic and syntactic thought are initially conceptualized
are unclear. For another, the illustrations he uses to
describe how gestures synchronize with important syntactic
breaks in spoken language are difficult to follow. Perhaps this
form of research, like studies of American Sign Language, can
only be adequately illustrated by a videotape and not by drawings.
Levelt's initial stage of conceptualization seems justified. After
all, speech does not start from nothing, and if it does not start with
concepts, how else could it possibly begin? At the same time, we
realize how difficult it is to actually define this stage in non-mentalistic
terms, and despite the plausibility of McNeill's binary model
of language and gestures being birthed together, like twins, it is
difficult to muster any hard evidence to support this, or any other
theory for the embryonic development of speech. Although we
know very little about how speech is initiated at this first stage of
conceptualization, we have psycholinguistic evidence to help us
understand the successive stages of production, so it is easier for us
to describe and to understand Levelt's second stage, formulation.
Formulation
Introduction
We have seen that the initial stage of conceptualization is so far
removed from the words we actually speak and write that it is
difficult to delineate this phase of production. But at the second
stage of speech production, formulation, we move close enough
to the eventual output of the process to allow us to be more
precise in our terminology and more convincing in our use of
empirical data. Conceptualization is hard to conceptualize, but
formulation is much easier to formulate. Well over three decades
ago, the psychologist Karl Lashley published one of the first
attempts to account for the way speakers sequence strings of
sounds, words, and phrases together so rapidly and accurately,
and his essay was influential enough to be included in the first
book ever published in English which focused exclusively on the
then very new field of the psychology of language. His essay was
first presented as an oral address, and it is intriguing to see how
Lashley organized it to demonstrate some of the very concepts
about speech production which he was writing about. For
example, he talked about how common it is to commit spelling
errors when one is typing, and he mentioned how he misspelled
‘wrapid’ with a w, while typing rapid writing', most probably
because as he was about to type 'rapid', he anticipated the 'silent w’
in the following word. These slips of the tongue, or pen, or computer
keyboard, are of keen interest to us in this chapter on production.
A moment later in his talk, to illustrate several of the
themes that were central to his presentation, Lashley gave the following
utterance as an example of how we comprehend spoken
sentences.
Rapid righting with his uninjured hand saved from loss the
contents of the capsized canoe
Unlike the first set of examples, these slips do not involve individual
sounds; rather, they seem to reflect a higher level of linguistic
organization because they are associated with complete
words, or with significant parts of words. (5) and (6) are very
common examples, and they remind us of the spoonerisms discussed
earlier. Notice how the misspoken forms still adhere to
normal patterns of word usage. For example 'Sesame Street
crackers' might be a brand of cracker named after the children's
TV show. Note too the manner in which (6) adheres to a regular
word pattern in English. A four-door sedan' has four doors, an
‘apple pie’ is made of apples. A common error by learners of
English is to call these objects a 'four-doors sedan' and 'apples
pie', following the logical, but non-English pattern of extending
the plural to the formation of noun phrases. But native speakers,
who follow the rules of word formation, do not simply swap the
two words that are reversed in (6) and say 'word of rules formation.
Even during the micromomentary process of formulating
their speech, they follow the regular and established pattern.
Examples (7) and (8) are further elaborations of this same
theme, but in this case, the suffix slots are exchanged while the
original words remain the same. The person who misspoke (7)
might have been thinking, if an 'American' is someone who lives
in America, why isn't a resident of New York a New Yorkan?"
And by the same logic, it 'arrival is the noun form of the verb
‘arrive’, why isn't the noun form of 'to derive', 'derival?' Once
again we witness the way slips of the tongue provide psycholinguistic
insights into the production of speech; they help us see
how speakers arrive at derivations.
Speech errors are also helpful in revealing a third level of language
processing at the formulation stage; they give support to
the notion that utterances are not just strings of sounds and linear
sequences of words, but are formed into larger structural units.
This is demonstrated in examples (9) and (10).
The affirmative tag is it' in (d) tells us that even though there is
no overt negation in this short response, it is still considered
grammatically negative. If it weren't, it would have a negative tag
at the end, wouldn't it? In contrast, both (e) and (f) must have
negative tags because they are both grammatically affirmative
phrases. Nevertheless, they are not similar: (e) expresses the concept
of 'not important' by choosing an antonym; (f) says the same
thing by using a negative prefix with 'important'. And the choice
of prefix in (f) is an additional complication, because English
proffers a wide range of potential prefixes. Some words take
several choices (for example unEuropean, non-European, anti- European),
but even words like important', which only take
un-, create a production problem. When speakers choose 'trivial',
the only thing they have to remember is the word itself, but when
they select a word which takes a negative prefix, they have to
recall which one to use. Again, we often fail to realize the complexity
of the production process until we see it fall apart, as in the
incorrect choice of a prefix by a learner (for example unpossible
and disimportant),
Let us go back now to the initial choice the speaker made.
Recall that the first alternative the speaker had in formulating the
concept not important', was whether to express the negative
response lexically-via words, or grammatically-via the use of
syntactic negation. Let us suppose the second alternative was
picked, creating the series of choices exemplified by sentences (g)
and (h).
You may have to glance at these twice to catch the slight differences
between them, and they are so minimal that you might be
provoked into using some of the earlier examples: the differences
are nothing; they're trivial! But if the speaker has chosen to
express the negation grammatically rather than through word
choice, important differences can be indicated by means of stress.
Normally, the contracted negative in (g) is chosen because negation
is typically not the focus of our attention, but (h) offers an
effective way of emphasizing negation. Supposing you are in an
argumentative state, and your conversational partner keeps
insisting that the situation is desperate; (h) allows you to be
emphatic about your denial. Put tersely, the difference between
the two sentences is that in (g) the negative is not usually stressed,
but in (h) it receives unusual stress.
The significance of these slight differences may seem minimal
within the context of the myriad sounds, words, and sentences
that comprise our daily staple of communication, but along With
the slip of the tongue examples, they demonstrate the enormous
number and intricacy of choices facing a speaker, or a writer, at
this important stage of formulation.
Articulation
We have spent considerable time examining the second stage of
speech production, and for good reason. Like the operation of a
computer program during word processing, the formulation
stage of speech involves thousands of split-second decisions
regarding the hierarchical and sequential selection of myriads of
potential segments. But this third stage of articulation is similar to
what happens when all of those bits of information selected by a
word processing program go from your computer to your printer;
unless this vast amount of electrical data is 'articulated' into
letters of the alphabet and successfully printed, no message is
received. In fact, if the printer is not functioning properly, there is
no evidence that the message was ever even composed. So, too,
with the production of speech. Unless all of the electrical impulses
streaming from your brain in the form of speech are transformed
into audible and comprehensible articulations, no words are
heard and nothing is communicated. The conceptualization stage
might pompously perceive itself as the primary and ultimate composer
of communication, and the formulation stage might pride
itself as the conductor and orchestrator of speech sounds, but
without the instruments of articulation, the music of our voices
remains unheard and unappreciated. Like the operations of a
printer or the playing of instruments then, the articulation of
speech sounds is a vital third stage of production and, quite naturally,
attracts the interest of psycholinguists.
As recently as the 1960s, linguists upheld the common sensical
and seemingly incontrovertible notion that the chest, throat, and
mouth were anatomical organs designed solely for biological
functions. Only in a secondary way could they be considered the
organs of speech. Surely, the basic function of our lungs is to
exchange oxygen for carbon dioxide, not to produce syllables
and, most assuredly, the primary use of our teeth is for chewing,
not for the articulation of sounds like [t] and [d]? True as these
assertions may be, they do not preclude the possibility that some
organs may have been shaped in their recent evolutionary history
to enhance the production of human speech sounds. Some thirty
years ago Eric Lenneberg, a psycholinguist, showed that whereas
the majority of these organs have primarily evolved to serve
essential biological functions such as respiration and ingestion, a
few of them have adopted secondary functions connected with
the enhancement of speech articulation. In a few cases, there are
organs that have changed anatomically to fit this new role as
speech articulator. So dramatic are the changes that they differ
physically from the corresponding organs in closely related
species like chimpanzees.
Perhaps the most dramatic example of an organ which has
adapted itself for human articulation is the larynx-the voice
box' which houses our vocal cords. Like all the other speech
organs, the larynx did not initially evolve with the specific function
of helping humans to articulate language. For one thing, the
vocal cords in all animals possessing a larynx serve as a kind of
emergency trap door which can prevent foreign matter, such as
bits of food, from falling from the mouth down the pharyngeal
tube and through the trachea into the lungs. When bits of debris
do manage to find their way down these passageways, the vocal
cords help control explosions of air from the lungs to cough this
potentially life-threatening jetsam back up out of the mouth. The
larynx thus helps keep the respiratory tract clear, but it serves
another primary function which is just the opposite of coughing.
By squeezing tightly closed, it can trap air in the chest cavity and
create a solid fulcrum for the limbs to work against when heavy
physical exertion is required. However, in the case of our species
we could claim that speech has now become the primary function
of the larynx and the other, original purposes of the voice box
have diminished to secondary stature.
Evidence for the evolutionary modification of the human larynx
to create speech is quite dramatic. Lenneberg and others have
documented several speech-enhancing characteristics of the voice
box that are unique to humans and are absent in other mammals,
even the primates like chimps and gorillas. No wonder then that
they have remained unable to master articulate speech. The most
striking difference between humans and all other animals in this
area of the body is the position of the larynx. In all other animals,
the larynx is found high in the throat, crammed behind the
tongue, an exceedingly advantageous position for preventing
debris from entering the trachea, for it can be trapped immediately
as it leaves the mouth. But this is not true for us. Feel your
neck and find your larynx (or Adam's apple'). You will locate
about halfway down, almost touching the top of a high-collared
shirt or blouse. Consider the awful consequences of this anatomical
deviation from natural evolution. Unlike all other animals,
our emergency trap door cannot stop foreign matter as soon as it
leaves the mouth. It is so far down that a passageway, called the
pharynx, has been created into which debris can easily fall, and
we, among all creatures, are the most susceptible to choking. Why
does nature deviate in this destructive manner for humans?
The advantages of the lower voice box reside in the way this
arrangement serves to embellish the articulation of speech
sounds. Unlike other mammals whose highly-positioned larynx
virtually precludes the existence of a pharyngeal tube linking the
back of the mouth with the opening of the vocal cords, the pharynx
benefits the production of speech in at least two ways. It creates
a new source of speech sounds-the throaty consonants of
Arabic, or the initial consonant of the two words in the English
salutation, Hi Harry!' A pharyngeal tube also increases resonance
by adding extra acoustic space to the already existing oral
and nasal cavities. The addition of a pharynx to the vocal repertoire
is not unlike the addition of a cello to the duet of a violin and
viola; the timbre of the human voice is that much richer, thanks to
the added instrument. Another enormous benefit of the lowered
larynx is the way it frees up the back of the tongue so that the
tongue root can maneuver and create more speech sounds. The
contrast between the vowels in the words look' and 'Luke' is
made largely by subtle movements of the tongue root, movements
that no other animals are capable of performing. So the linguistic
advantages outweigh the physiological disadvantages, and if the
emergence of language is as vital to our evolutionary history as
most anthropologists believe, and if language is so indispensable
to our species, it is no exaggeration to claim that the descent of the
larynx has permitted the ascent of mankind!
Given the anatomy of articulation we have been endowed with,
what do we know about the programming of articulation? How
do sounds trip so miraculously off the tips of our tongues once
speech is conceptualized and formulated? It is easy to assume that
speech sounds are produced in a linear, sequential fashion, like
cars off an assembly line, but a closer analogy might be to the
team effort that goes into producing a batch of cookies. While one
person might be chopping the walnuts, another might be preparing
the cookie dough, while a third might be preheating the oven
and greasing the cookie sheets. So too in the production of speech
sounds. The process might appear to be linear, but the lungs, larynx,
and lips may be working all at the same time, and coarticulation
is the norm, not the exception. That is, in the production of
any single sound, a lot of anatomical effort is devoted to performing
several different movements simultaneously.
Consider just one sound, the second [k] (spelled, of course,
with a 'qu) in the expression: Keep quiet kid!' Let us begin by
contrasting the second [k] with the first (in 'keep) and the last (in
kid'). All three of these sounds are dorsovelar which means they
are made in the back of the mouth. The back of the tongue (the
dorsum) hunches up and touches the soft palate at the back of the
mouth (the velum) to stop the flow of air momentarily to produce
the consonant [k]. But the process is much more complicated than
this, because every sound in the stream of speech is affected by the
sounds which swim around it. None of the [k]s in this particular
phrase is preceded by any sound which would demonstrably
affect its pronunciation but each is followed by sounds which do
affect articulation. 1The first and last are followed by a vowel pronounced
in the front of the mouth, so the [k] in 'keep and the [k]
in 'kid' slide forward a bit from their usual position in the back. It
is almost as if the tongue were at the starting line of a sprint and
was trying to inch up a bit to get a head start in the race to those
front vowels. There is a double contrast between the [k] in 'quiet
and the first and the last [k]'s. First, because the vowel in this
word at least begins with a sound in the back of the mouth (the
initial [a] of the diphthong [ai]), the [k] in this word does not inch
forward at all, but actually sits well back in order to hit the following
back vowel. Secondly, because 'qu' is actually a cluster
[kw] and not a single sound, the initial [k] is rounded in anticipation
of the following [w]. In other words, the lips assume an ‘o’
position when we begin to articulate ‘quiet’, whereas they
remained flat and unrounded in the production of the first and
last words. Try reproducing this phrase in front of the mirror, and
you'll obtain visual evidence of the effects of coarticulation; the
mouth momentarily puckers up when you begin to pronounce
‘quiet’. Here we observe the complexity of articulation. Sounds
do not emerge as segments strung together sequentially; they are
mixed and melded, with each sound shaping its neighbours while
concurrently being shaped themselves.
Psycholinguists have developed a number of competing models
to try to account for the complexity of speech articulation, and
they have tried to employ various sources of evidence to peek into
this complicated process, but much of articulation remains a
mystery. For example, despite the increasing sophistication of
modern neurology and the development of techniques such as
Positron Emission Tomography (PET) Scans to examine the way the
human brain programs neuromuscular movements, we still have
little understanding of how the cerebral software programs the
anatomical printer to articulate sounds in such a glib manner. Let
us narrow the issue down to one simple question. How does the
tongue 'know' to cheat a little bit ahead in the first and last words
of the phrase described in the paragraph above, but ‘know’ not to
inch forward in the second word. And how do the lips ‘know’
when to pucker up? It would be impossibly difficult to explain the
rapidity and accuracy of articulation in such closely related
phrases simply as a chain of habits acquired in a linear way.
We see then that even at this seemingly uncreative and
mechanical aspect of speech production, complexity and
mystery abound, and speaking ceases to appear a simple and
mundane act. Speech production does not end with articulation,
however the fourth and final stage of production is the process of
self-monitoring.
Self-monitoring
Earlier, during our review of slips of the tongue, it was noted that
the production process sometimes goes awry and speakers will
verbally misstep, especially with irregular or more unusual
forms. Almost always, however, they instantly catch themselves,
retreat a step, and correctly recreate the intended sequence, as in
(1) and (2).
(1) The last I knowed about it (I mean knew about it), he
had left Vancouver.
(2) She was so drank {I mean drunk), that we decided to
drive her home.
Comprehension: understanding
what we hear and read
Notice that in every case, the subjects heard eel as the key word
in the sentence, but most of the subjects claimed they had heard a
different word for each example-specifically, wheel for (1), heel
for (2), peel for (3), and meal for (4). The insertion of a different
missing sound (phoneme) to create a separate but appropriate
‘eel’ word in each sentence is called the phoneme restoration effect.
Under these conditions, listeners do not accurately record
what they hear; rather, they report what they expected to hear
from the context, even if it means they must add a sound that was
never actually spoken at the beginning of the target word. Several
simple but significant observations can be drawn from this
sample of the early psycholinguistic research into the nature of
comprehension.
First of all, as just illustrated, people don't necessarily hear each
of the words spoken to them. Comprehension is not the passive
recording of whatever is heard or seen; listeners are not tape
recorders nor readers video cameras. Second, comprehension is
strongly influenced by even the slightest of changes in discourse
which the listener is attending to. In these examples, except for
the last word, each of these sentences is identical. Finally, comprehension
is not a simple item-by-item analysis of words in a linear
sequence. We don't read or hear the same way we count digits
sequentially from one to ten. Listeners and readers process
chunks of information and sometimes wait to make decisions on
what is comprehended until much later in the sequence. It is the
last-not the sixth or ‘target’-word in each of the four examples
above which dictated what the listeners in the experiment
reported they heard. We don't seem to listen to each word individually
and comprehend its meaning in isolation; we seek contextual
consistency and plausibility, even if it comes to adding a
sound or inventing a word that wasn't actually spoken. This
chapter then reviews some of the ways in which psycholinguistic
processes affect the way listeners and readers comprehend
language.
Although, in the course of everyday conversation, we don't
hear vowels and consonants as isolated sounds, we can, with the
help of machines, measure acoustic information extremely precisely.
The /p/ in the following English words is pronounced
slightly differently depending on where it occurs in the word or
what other sounds follow it. The initial /p/ of 'pool' is pronounced
with puckered lips but the 'same' /p/ in ‘peel’ is spoken
with the lips spread, and neither of these /p/'s sound quite like the
/p/ in 'spring. Although these details may seem trivial to a native
speaker of English, they are significant enough acoustically to be
heard as contrasting phonemes in other languages. Despite these
differences, and other variations of /p/ that could be cited in
countless other examples, native speakers of English claim they
hear and pronounce the same /p/ sound. Notice that for these and
most of the other examples, we spell the sound with the letter ‘p’
and furthermore, despite all the variations in /p/, native speakers
of English almost never confuse any manifestation of the /p/
sound with /b/, which is acoustically very similar. Recall that in
the discussion of the articulation stage in Chapter 3, we saw that
there was a sizeable phonetic difference between the initial /k/s of
‘keep’ and kid' and the /k/ sound which begins the word 'cool'.
Phoneticians have been fairly successful in writing rules that predict
which precise acoustic form of /p/ is pronounced (or heard)
under which phonetic condition; nevertheless, they have been
unable to explain how this variation is processed by the mind or
how all the phonetic differences which occur among all the many
languages of the world can be accounted for in terms of the common,
universal processes of perception that are shared by all
humans. Although the exact details of this acoustic processing
have yet to be resolved, psycholinguists have come up with some
explanations for this most fundamental level of comprehension
Suppose we are engaged in conversation with a friend and are
discussing two other acquaintances with similar sounding
names ‘Benny’ and ‘Penny’. What phonetic information do we
employ as we listen to distinguish these names which are identical
in pronunciation except for the initial consonant? Phoneticians
have discovered that the main feature which English speakers
attend to, albeit unconsciously, is the Voice Onset Timing (VOT) of
the initial consonant. Using instruments which are sensitive
enough to measure contrasts as small as milliseconds in the duration
of speech sounds, they have demonstrated that the most significant
acoustic difference between English consonants like /b/
and /p/ is the length of time it takes between the initial puff of air
that begins these sounds, and the onset of voicing in the throat
that initiates any vowel sound which follows the consonants.
Since almost all the other phonetic features of this consonantal
pair are identical, the crucial clue that separates the voiced /b/ and
its voiceless counterpart /p/ is a VOT of a scant 5o milliseconds
This means that the correct comprehension of the name 'Penny',
as opposed to the mistaken recognition of the similar sounding
‘Benny’, depends on an ability to perceive a voicing delay of one
twentieth of a second! The simple task of recognizing which per-
son is being referred to during a conversation is based on your
ability to isolate one subtle phonological feature from the myriad
sounds hitting your ear and to make a split-second judgment,
How do speakers of English, or any language for that matter,
make these incredibly difficult decisions about speech so rapidly
and so accurately?
It appears that the acquisition of this phonetic ability cannot be
completely explained only by exposure to, or instruction in, the
language. In other words, native speakers do not acquire all of
this acoustic information from direct experience with language,
and as we learned in Chapter 2, parents and caretakers do not
provide explicit instruction on these matters. Even phoneticians
do not subject their children to hours of nursery training listening
to minimal pairs like 'pie' versus 'buy'. Psycholinguists have discovered
through careful experimentation that humans are actually
born with the ability to focus in on VOT differences in the
speech sounds they hear, and they have proven that rather than
perceiving VOT contrasts as a continuum, people tend to categorize
these minute phonetic differences in a non-continual,
binary fashion.
All of this has been decisively documented in experiments
where native speakers of English listened to artificially created
consonant sounds with gradually lengthening VOTs and were
asked to judge whether the syllables they heard began with a
voiced consonant (like /b/ which has a short VOT) and a voiceless
one (like /p/ which, as was just pointed out, has a VOT lag of
about 50 milliseconds). When subjects heard sounds with a VOT
of about 25 milliseconds, about halfway between a /b/ and a /p/,
they rarely judged the sound to be 5o% voiceless and so%
voiced, they classified it as one sound or the other. This phenomenon
is called categorical perception. Psycholinguists have been
able to prove the presence of categorical perception in very young
infants, through a series of cleverly designed experiments. And in
equally ingenious research with several species of animals, they
have found, by and large, that this kind of all-or-nothing acoustic
perception does not exist in other species, Categorical perception
is seemingly unique to human beings, and appears to qualify as
one aspect of universal grammar (UG), the genetic propensity for
comprehending and producing language which most psycholinguists
believe is a uniquely human endowment. These experiments
with VOT perception in human infants are one of the few
solid pieces of evidence we have that UG exists and that at least
part of human language is modular-that is, some parts of language
reside in the mind or brain as an independent system or
module.
Although categorical perception of VOT is an ability children
are born with, it is also influenced by the linguistic environment
a child is raised in. Here lies the second part of the puzzle of
how native speakers of English grow up with the intrinsic ability
to distinguish instantly between the names 'Benny' and Penny'.
Because the English language divides the VOT spectrum into two
sets of sounds, for example the voiced and voiceless pairs of consonants
/b/ versus /p/, /d/ versus /t, and /g/ versus /k/, children
learning English acquire the ability to use their innately specified
gift of categorical perception to divide the VOT continuum into
two equal halves, corresponding to the voiced and voiceless consonants
just exemplified. On the other hand, children exposed to
a different language, say Thai, which has three, not two, VOT
consonantal contrasts, grow up after years of exposure with
the ability to make a three-way categorical split. Thus Thai
children rapidly acquire the ability to hear an extremely short
VOT as /b/ (as in /bail, the Thai word for leaf), a slightly
longer VOT as /p/ (a sound like the /p/ in the English word
'spring', as in /pai/, the Thai word for 'go'), and any VOT longer
than 50 milliseconds as an aspirated /ph/ (a sound very close
to the English /p/ and which is used in the Thai word, /phai/,
which means 'paddle).
When any language learner, whether a child learning their first
language, or an adult a second language, is exposed to the VOT
settings of a particular language over an extended period of time
with lots of opportunities for acoustic input, it appears that they
use their innate ability to hear speech sounds categorically to
acquire the appropriate VOT settings. The successful comprehension
of speech sounds is, therefore, a combination of the innate
ability to recognize fine distinctions between speech sounds
which all humans appear to possess, along with the ability all
learners have to adjust their acoustic categories to the parameters
of the language, or languages, they have been immersed in. We see
then that learning to comprehend, like all aspects of language
acquisition, is again a merger of both nature and nurture.
(11) It's not true that Wednesday never comes after a day
that's not Tuesday.
With hocked gems financing him, our hero bravely defied all
scornful laughter that tried to prevent his scheme. Your eyes
deceive you, he had said, an egg not a table correctly typifies
this unexplored planet. Now three sturdy sisters sought
proof, forging along sometimes through calm vastness, yet
more often over turbulent peaks and valleys. Days became
weeks as many doubters spread fearful rumors about the
edge. At last, from nowhere, welcome winged creatures
appeared, signifying momentous success.
Comprehension concluded
Once again we have discovered that an everyday activity that
seems to be simple and straightforward is, upon more intensive
scrutiny, complex and variable. In the comprehension of speech
sounds, we see further evidence that some parts of human language
are innate, and do not have to be learned. The perception of
major linguistic differences in sounds, such as VOT, is hard-wired
into the human brain, and even young children demonstrate the
ability to classify very small differences in VOT into one or
another phonetic category. This innate ability is extremely useful
for children as they grow up hearing their mother tongue, because
it allows them to pick up the few significant differences in that
particular language and, at the same time, to ignore the many
which are insignificant.
The research into the comprehension of words has shown that
we are very much affected by context, and that our understanding
is both facilitated and complicated by the different pieces
of knowledge we possess for each logogen. It is clear from the
TOT phenomenon that we have access to a dictionary-like memory
for words. We can 'search' for a partially-remembered word
by comparing and contrasting other words which share similar
specifications. But our knowledge of and about words is much
more extensive: the meaning of a word immediately triggers a
spreading activation of associations which help us understand it
in many different contexts, and may bring other related words to
mind.
The grammatical structure of a sentence might initially influence
the garden path we choose in trying to understand it, but the
greatest influence on sentence comprehension is meaning. We can
see this in the experiments with ambiguous sentences because it is
clear that ambiguity slows down processing time, but we also
observe it in recall. People remember the 'what' that is spoken or
written better than the 'how'. Finally, comprehension of larger
units of language also indicates the importance of meaning. Texts
that fit into a context which we understand and expect are com
prehended more quickly and remembered more readily than ones
which are presented to us without a context.
It is plain that only a complex model of comprehension like
PDP can begin to account for the way readers and listeners comprehend
the millions of linguistic messages they receive each day.
Psycholinguists have to develop a model of comprehension that
successfully integrates all the diverse, yet parallel and simultaneous
processes that we have examined in this chapter, and
obviously such a model will be exceedingly elaborate. It will have
to begin with the innate mechanisms for language that are wired
into the human mind. It will have to account for the way in which
young children rapidly learn to extricate significant phonemes,
words, sentence structures, and phrases from the multitude of
sounds and sights that besiege them each day. And ultimately it
will have to explain how some sort of executive decision-maker in
the mysterious garden of the human mind decides when to continue
along one path toward understanding, when to abandon it
abruptly for a more fruitful alternative, and how to seek almost
always successfully an accurate interpretation of the intended
message.
Broca's aphasia
[The patient is attempting to describe an appointment for
dental surgery.]
Yes... ah… Monday ... er ... Dad and Peter H ..., and Dad…
er ... hospital... and ah... Wednesday… Wednesday, nine
o'clock … and oh...Thursday ... ten o'clock, ah doctors.
two... an' doctors.. and er... teeth... yah
Wernicke's aphasia
[The patient is trying to describe a picture of a family in a
kitchen.]
Well this is... mother is away here working her work out
o'here to get her better but when she's looking, the two boys
looking in the other part. One their small tile into her time
here. She's working another time because she's getting too…
Concluding summary
What do all these examples of speech dissolution tell us about the
nature of language and mind? Well, for one thing, given the unbelievable
complexity of human language, it is quite astounding to
realize that among the world's more than five billion speakers,
only a remarkably small number of them are afflicted with any of
the communicative anomalies reviewed in this chapter. When we
consider the intricacy of acquisition, production, and comprehension
involved in just one language, our mother tongue, and then
add to this the fact that nearly half the world’s population are
bidialectal if not bilingual, and are able to process two distinct
varieties of language successfully, it is amazing that dissolution is
a comparative rarity and not the norm. So the first thing we learn
from all of these studies of aberrant language is that because they
are abnormal, the everyday use of language without disorders in
acquisition, production, or comprehension is a wonder of
miraculous proportions.
Second, we can acknowledge from the neurological examples
which were reviewed in this chapter that there is strong evidence,
from the way the brain processes information, for the unique
independence of language. In all varieties of aphasia and in many
of the neurolinguistic studies of patients who have undergone
major brain surgery, it is plain that language and speech enjoy
a unique neurological status in the human brain, and we find
support for the notion that the capacity to comprehend and
produce language is hard-wired to the mid-central area of the left
hemisphere for most adults. At the same time, evidence was
presented to indicate that speech and language are not always
narrowly and immutably localized to one area of the brain. For
young children especially, language seems to be more diffusely
controlled by both hemispheres. Indeed, one area of neurolinguistics
that needs to be more fully examined is how and why language
shifts from a broader, bilateral representation in young
children to a narrower, unilateral control in adolescents and
adults. An even more intriguing puzzle remaining to be solved is
why the neurolinguistic evidence tends to support the independence
of speech and language from other aspects of behavior,
whereas the psycholinguistic data suggests just the opposite-
that language is part and parcel of cognition and perception.
When we turn to examples of dissolution that do not seem to be
caused by brain damage, we discover that the data from research
on speech and hearing disorders does not differ significantly from
the information we have on normal development. The study of
stuttering, for example, endorses the notion that the formulation
stage is an important level of speech production. But in general,
all of these disabilities, irrespective of their origins, whether
behavioral, as in stuttering, or clearly genetic, as in DS, or the
natural forces of maturation, as in aging, or due to a still
unknown combination of forces, such as autism, point to the
third and most significant conclusion. By and large, language
seems to be closely related to other aspects of human behavior,
particularly to cognition.
In summary, the disruptions in the environment or in the
genetic code that bring about speech and language disabilities
never seem to single out language: they affect linguistic communication
because they afflict cognition and perception as a whole.
For this reason, psycholinguistics is drawn by language into a
more general inquiry of the workings of the human mind.