Tone and Voice - A Derivation of The Rules of Voice-Leading From Perceptual Principles PDF
Tone and Voice - A Derivation of The Rules of Voice-Leading From Perceptual Principles PDF
Tone and Voice - A Derivation of The Rules of Voice-Leading From Perceptual Principles PDF
Music Perception
Fall 2001, Vol. 19, No. 1, 164
Tone and Voice: A Derivation of the Rules of VoiceLeading from Perceptual Principles
D AV I D H U R O N
N the training of Western musicians, the canon of harmony and voiceleading rules has been considered one of the essential foundations of the
art-music craft. Of course not all composers accede to the norms of traditional harmony or voice-leading. Nor should they. Rules ought to be followed only when the composer agrees that the goal or goals embodied by
the rules form worthy musical objectives, and when the rules themselves
provide an effective means by which the goal(s) may be achieved. We can,
therefore, ask three questions of any procedural rule: (1) What goal is served
by following the rule? (2) Is the goal worthwhile? and (3) Is the rule an
effective way of achieving the purported goal?
The practice of Western harmony and voice-leading has been the subject
of extensive theoretical attention and elucidation (Aldwell & Schachter,
Address correspondence to David Huron, School of Music, Ohio State University, 1866
College Rd., Columbus, OH 43210.
ISSN: 0730-7829. Send requests for permission to reprint to Rights and Permissions,
University of California Press, 2000 Center St., Ste. 303, Berkeley, CA 94704-1223.
1
David Huron
1989; Berardi, 1681; Ftis, 1840; Fux, 1725; Hindemith, 1944; Horwood,
1948; Keys, 1961; Morris, 1946; Parncutt, 1989; Piston, 1978; Rameau,
1722; Riemann, 1903; Schenker, 1906; Schoenberg, 1911/1978; Stainer,
1878; and many others1). Over the centuries, theorists have generated a
wealth of lucid insights pertaining to harmony and voice-leading. Earlier
theorists tended to view the voice-leading canon as a set of fixed or inviolable universals. More recently, theorists have suggested that the voice-leading canon can be usefully regarded as descriptive of a particular, historically encapsulated, musical convention. A number of twentieth-century
theorists have endeavored to build harmonic and structural theories on the
foundation of voice-leading.
Traditionally, the so-called rules of harmony are divided into two broad
groups: (1) the rules of harmonic progression and (2) the rules of voiceleading. The first group of rules pertains to the choice of chords, including
the overall harmonic plan of a work, the placement and formation of cadences, and the moment-to-moment succession of individual chords. The
second group of rules pertains to the manner in which individual parts or
voices move from tone to tone in successive sonorities. The term voiceleading originates from the German Stimmfhrung, and refers to the direction of movement for tones within a single part or voice. A number of
theorists have suggested that the principal purpose of voice-leading is to
create perceptually independent musical lines. This article develops a detailed exposition in support of this view.
In this article, the focus is exclusively on the practice of voice-leading.
No attempt will be made here to account for the rules of harmonic progression. The goal is to explain voice-leading practice by using perceptual principles, predominantly principles associated with the theory of auditory
stream segregation (Bregman, 1990; McAdams & Bregman, 1979; van
Noorden, 1975; Wright, 1986; Wright & Bregman, 1987). As will be seen,
this approach provides an especially strong account of Western voice-leading practice. Indeed, in the middle section of this article, a derivation of the
traditional rules of voice-leading will be given. At the end of the article, a
cognitive theory of the aesthetic origins of voice-leading will be proposed.
Understood in its broader sense, voice-leading also connotes the dynamic quality of tones leading somewhere. A full account of voice-leading
would entail a psychological explanation of how melodic expectations and
implications arise and would also explain the phenomenal experiences of
tension and resolution. These issues have been addressed periodically
throughout the history of music theory and remain the subject of continu1. No list of citations here could do justice to the volume and breadth of writings on this
subject. This list is intended to be illustrative by including both major and minor writers,
composers and noncomposers, dydactic and descriptive authors, different periods and nationalities.
ing investigation and theorizing (e.g., Narmour, 1991, 1992). Many fundamental issues remain unresolved (see von Hippel & Huron, 2000), and so
it would appear to be premature to attempt to explain the expectational
aspects of tone successions. Consequently, the focus in this article will be
limited to the conventional rules of voice-leading.
The purpose of this article is to address the three questions posed earlier,
namely, to identify the goals of voice-leading, to show how following the
traditional voice-leading rules contributes to the achievement of these goals,
and to propose a cognitive explanation for why the goals might be deemed
worthwhile in the first place.
Many of the points demonstrated in this work will seem trivial or intuitively obvious to experienced musicians or music theorists. It is an unfortunate consequence of the pursuit of rigor that an analysis can take an
inordinate amount of time to arrive at a conclusion that is utterly expected.
As in scientific and philosophical endeavors generally, attempts to explicate common sense will have an air of futility to those for whom the
origins of intuition are not problematic. However, such detailed studies
can unravel the specific origins from which intuitions arise, contribute to a
more precise formulation of existing knowledge, and expose inconsistencies worthy of further study.
Plan
My presentation will proceed according to the following plan. First, the
principal rules of traditional voice-leading will be reviewed. Following this
review, 10 principles of auditory perception will be described and their
musical implications identified. The perceptual principles described will
initially be limited to a core set of 6 principles that are most pertinent to
understanding voice-leading. These principles can be treated in a manner
akin to axioms in a formal system from which a set of propositions can be
derived.
In Part II, we will attempt to account for the traditional rules of voiceleading. In the process of the derivation, several novel rules will arise that
are not normally found in theoretical writings on voice-leading. These novel
rules can be regarded as theory-derived predictions. For these novel rules,
we will examine compositional practice to determine whether composers
typically write in a manner consistent with these additional rules. To foreshadow the results, we will see that voice-leading practices are indeed consistent with several nontraditional rules predicted by the perceptual principles.
Part III presents four additional perceptual principles that are occasionally linked to the practice of Western voice-leading. In essence, these auxil-
David Huron
iary principles constitute options that shape the music-making in perceptually unique ways. Composers can elect to include or exclude one or
more of these auxiliary principles depending on the implied perceptual goal.
Once again, these principles can be treated in a manner akin to axioms in
formal logic and selectively added to the core group of six central principles. Depending on the choice of auxiliary principles, we will see that
alternative voice-leading systems arise. It will be argued that the choice of
such auxiliary principles is one of the hallmarks of musical genres; these
auxiliary principles shed light on such matters as the distinction between
homophonic and polyphonic voice-leading and such unique genres as close
harmony (such as the voice-leading found in barbershop quartets).
In the concluding section, a psychological account will be advanced whose
goal is to explain why conventional voice-leading might be experienced by
many listeners as pleasing. A number of testable predictions arise from this
account. The article closes by identifying a number of unresolved issues
and posing questions for further research.
Note that the purpose of this article is not somehow to vindicate or
otherwise act as an apologist for the traditional rules of Western harmony.
Nor is the intent to restrict in any way the creative enterprise of musical
composition. If a composer chooses a particular goal (either implicitly or
explicitly), then there are frequently natural consequences that may constrain the music making in such a way as to make the goal achievable.
Musically pertinent goals might include social, political, historical, formal,
perceptual, emotional, cognitive, and/or other objectives. The purpose of
this article is merely to identify and clarify some of the perceptual and
cognitive aspects that shape music making. In pursuing this analysis, readers should not assume that other musically pertinent goals are somehow
less important or irrelevant in understanding music.
pears in nearly every pedagogical work on voice-leading. These core voiceleading rules include the following:
1.
David Huron
noise bands are widely separated and masking is absent, listeners are much
less adept at identifying the number of noise bands present in a sound field
than an equivalent number of pure tonesprovided the tones are not
harmonically related (see below).
In the second instance, certain sets of pure tones may coalesce to form a
single auditory imageas in the case of the perception of a complex tone.
For complex tones, the perceptual image is strongly associated with the
pitch evoked by a set of spectral components. One of the best demonstrations of this association is to be found in phenomenon of residue pitch
(Schouten, 1940). In the case of a residue pitch, a complex harmonic tone
whose fundamental is absent nevertheless evokes a single notable pitch
corresponding roughly to the missing fundamental.
In the third instance, tones having inharmonic partials tend to evoke
more diffuse auditory images. Inharmonic partials are more likely to be
resolved as independent tones, and so the spectral components are less apt
to cohere and be perceived as a single sound. In a related way, inharmonic
tones also tend to evoke competing pitch perceptions, such as in the case of
bells. The least ambiguous pitches are evoked by complex tones whose
partials most closely approximate a harmonic series.
In general, pitched sounds are easier to identify as independent sound
sources, and sounds that produce the strongest and least ambiguous pitches
produce the clearest auditory images. Pitch provides a convenient hanger
on which to hang a set of spectral components and to attribute them to a
single acoustic source. Although pitch is typically regarded as a subjective
impression of highness or lowness, the more fundamental feature of the
phenomenon of pitch is that it provides a useful internal subjective label
a perceptual handlethat represents a collection of partials likely to have
been evoked by a single physical source. Inharmonic tones and noises can
also evoke auditory images, but they are typically more diffuse or ambiguous.
Psychoacousticians have used the term tonality to refer to the clarity of
pitch perceptions (ANSI, 1973); however, this choice of terms is musically
unfortunate. Inspired by research on pitch strength or pitch salience carried out by Terhardt and others, Parncutt (1989) proposed the term
tonalness; we will use the less ambiguous term toneness to refer to the
clarity of pitch perceptions. For pure tones, toneness is known to change
with frequency. Frequencies above about 5000 Hz tend to sound like indistinct sizzles devoid of pitch (Attneave & Olson, 1971; Ohgushi & Hato,
1989; Semal & Demany, 1990).2 Similarly, very low frequencies sound like
2. More precisely, pure tones above 5 kHz appear to be devoid of pitch chroma, but
continue to evoke the perception of pitch height. See Semal and Demany (1990) for further
discussion.
David Huron
Fig. 1. Changes of maximum pitch weight versus pitch for complex tones from various
natural and artificial sources; calculated according to the method described by Terhardt,
Stoll, and Seewann (1982a, 1982b). Solid line: pitch weight for tones from C1 to C 7 having
a sawtooth waveform (all fundamentals at 60 dB SPL). Dotted lines: pitch weights for
recorded tones spanning the entire ranges for harp, violin, flute, trumpet, violoncello, and
contrabassoon. Pitch weights were calculated for each tone where the most intense partial
was set at 60 dB. For tones below C6, virtual pitch weight is greater than spectral pitch
weight; above C6, spectral pitch predominateshence the abrupt change in slope. The figure shows that changes in spectral content have little effect on the region of maximum pitch
weight. Note that tones having a pitch weight greater than one unit range between about E2
and G 5corresponding to the pitch range spanned by the treble and bass staves. Recorded
tones from Opolko and Wapnick (1989). Spectral analyses kindly provided by Gregory
Sandell (Sandell, 1991a).
shows the calculated pitch weight of the most prominent pitch for sawtooth
tones having a 60 dB SPL fundamental and ranging from C1 to C7. The
dotted curves show changes of calculated pitch weight for recorded tones
spanning the entire ranges for several orchestral instruments including harp,
violin, flute, trumpet, violoncello, and contrabassoon. Although spectral
content influences virtual pitch weight, Figure 1 shows that the region of
maximum pitch weight for complex tones remains quite stable. Notice that
complex tones having a pitch weight greater than one unit on Terhardts
scale range between about E2 and G5 (or about 80 to 800 Hz). Note that
this range coincides very well with the range spanned by the bass and treble
staves in Western music.
The origin of the point of maximum virtual pitch weight remains unknown. Terhardt has suggested that pitch perception is a learned phenomenon arising primarily from exposure to harmonic complex tones produced
typically by the human voice. The point of maximum virtual pitch weight
is similar to the mean fundamental frequency for women and childrens
voices. The question of whether the ear has adapted to the voice or the
voice has adapted to the ear is difficult to answer. For our purposes, we
might merely note that voice and ear appear to be well co-adapted with
respect to pitch.
In Huron and Parncutt (1992), we calculated the average notated pitch
in a large sample of notes drawn from various musical works, including a
large diverse sample of Western instrumental music, as well as non-Western works including multipart Korean and Sino-Japanese instrumental
works. The average pitch in this sample was found to lie near D 4a little
more than a semitone above the center of typical maxima for virtual pitch
weight. This coincidence is especially evident in Figure 2, where the average notated pitch is plotted with respect to three scales: frequency, log frequency, and virtual pitch weight. The virtual pitch weight scale shown in
Figure 2 was created by integrating the curve for the sawtooth wave plotted in Figure 1that is, spreading the area under this curve evenly along
the horizontal axis. For each of the three scales shown in Figure 2, the
lowest (F2) and highest (G5) pitches used in typical voice-leading are also
plotted. As can be seen, musical practice tends to span precisely the region
of maximum virtual pitch weight. This relationship is consistent with the
view that clear auditory images are sought in music making.
Once again, the causal relationship is difficult to establish. Musical practice may have adapted to human hearing, or musical practice may have
contributed to the shaping of human pitch sensitivities. Whatever the causal
relationship, we can note that musical practice and human hearing appear
to be well co-adapted. In short, middle C truly is near the middle of
something. Although the spectral dominance region is centered more than
an octave away, middle C is very close to the center of the region of virtual
10
David Huron
Fig. 2. Relationship of pitch to frequency, log frequency, and virtual pitch weight scales.
Each line plots F2 (bottom of the bass staff), C4 (middle C), and G5 (top of the treble staff)
with respect to frequency, log frequency, and virtual pitch weight (VPW)-scale. Lines are
drawn to scale, with the left and right ends corresponding to 30 Hz and 15,000 Hz, respectively. The solid dot indicates the mean pitch (approx. D#4) of a cross-cultural sample of
notated music determined by Huron and Parncutt (1992). The VPW scale was determined
by integrating the area under the curve for the sawtooth wave plotted in Figure 1. Musical
practice conforms best to the VPW scale with middle C positioned near the center of the
region of greatest pitch sensitivity for complex tones.
pitch sensitivity for complex tones. Moreover, the typical range for voiceleading (F2-G5) spans the greater part of the range where virtual pitch weight
is high. By contrast, Figure 2 shows that the pitch range for music-making
constitutes only a small subset of the available linear and log frequency
ranges.
Drawing on the extant research concerning pitch perception, we might
formulate the following principle:
1. Toneness Principle. Strong auditory images are evoked when tones
exhibit a high degree of toneness. A useful measure of toneness is provided by virtual pitch weight. Tones having the highest virtual pitch
weights are harmonic complex tones centered in the region between F2
and G5. Tones having inharmonic partials produce competing virtual
pitch perceptions, and so evoke more diffuse auditory images.
11
most obvious factor is temporal continuity, in which sound energy is maintained over a period of time.
Auditory images may be evoked by either real (sensory) or imagined
(purely mental) processes. Two examples of purely mental auditory images
can be found in echoic memory and auditory induction (Houtgast, 1971,
1972, 1973; Thurlow, 1957; Warren, Obusek, & Ackroff, 1972).3 In the
case of auditory induction, Warren et al. generated stimuli in which intermittent faint sounds were alternated with louder sounds. The faint and
loud sounds were contiguous but not overlapping. Nevertheless, the faint
sounds tend to be perceived as a continuous background tone against which
the loud sounds are heard to pulse. In this case, there is a mental disposition to continue a sound image even when it is physically absent.
Warren et al. were able to explain the origin of this phenomenon by
showing that the frequency/intensity thresholds for auditory induction coincide closely with the thresholds for auditory masking. In other words,
auditory induction mentally reinstates sounds that the listener would expect to be maskedeven when the sounds are truly absent. With pure tones,
robust auditory induction effects can be achieved for durations of up to
300 ms. In the case of noise bands, auditory induction may be achieved for
durations of 20 s or more.
Auditory induction may be viewed as an involuntary form of auditory
imagination. Fortunately, auditory induction is a subjective phenomenon
that readily admits to empirical investigation and measurement. Other pertinent subjective experiences are not so easily studied. Listeners (musicians
especially) are also cognizant of the existence of voluntary auditory imagination, in which the listener is able to form or sustain a purely mental
image of some sound, such as the imagined sound of a tympani roll or the
sound of a bubbling brook.
Although imagined sounds may be quite striking, in general, imagined
sounds are significantly less vivid than actual sound stimuli. Moreover,
even in the absence of sound, recently heard sounds are more vivid than
less recently heard sounds. In short, auditory images have a tendency to
linger beyond the physical cessation of the stimulus. In the case of real
sounds, the evoked auditory images decay through the inexorable degradation of the short-term auditory storewhat Neisser (1967) dubbed echoic
memory. In general, the longer a sound stimulus is absent, the less vivid is
its evoked image.
A number of experiments have attempted to estimate the duration of
echoic memory. These measures range from less than 1 s (Treisman &
Howarth, 1959) to less than 5 s (Glucksberg & Cowen, 1970). Typical
3. Thurlow referred to this phenomenon as the auditory figure-ground effect; Houtgast
used the term continuity effect. We have elected to use the terminology proposed by
Warren et al.
12
David Huron
measures lie near 1 s in duration (Crowder, 1969; Guttman & Julesz, 1963;
Rostron, 1974; Treisman, 1964; Treisman & Rostron, 1972). Kubovy and
Howard (1976) have concluded that the lower bound for the half-life of
echoic memory is about 1 s. Using very short tones (40-ms duration), van
Noorden (1975, p. 29) found that the sense of temporal continuation of
events degrades gradually as the interonset interval between tones increases
beyond 800 ms.
On the basis of this research, we may conclude that vivid auditory images are evoked best by sounds that are either continuous or broken only
by very brief interruptions.4 In short, sustained and recurring sound events
are better able to maintain auditory images than are brief intermittent stimuli.
Drawing on the preceding literature, we might formulate the following
principle:
2. Principle of Temporal Continuity. In order to evoke strong auditory
streams, use continuous or recurring rather than brief or intermittent
sound sources. Intermittent sounds should be separated by no more
than roughly 800 ms of silence in order to ensure the perception of
continuity.
13
14
David Huron
the damping of physical vibrators is consistent with the goal of maintaining the illusion of a single continuous acoustic activity.
Of course music making also entails the use of brief sounds, such as
sounds arising from various percussion instruments like the wood block.
However, musicians tend to treat such brief sounds differently. Brief sounds
tend not to be used to construct lines of soundsuch as melodies
instead, these sounds are typically used intermittently. Those percussion
instruments having the shortest tone durations are least apt to be used for
melodic purposes. When instruments having rapid decays are used to perform melodies, they often employ tremolo or multiple repeated attacks, as
in music for marimba or steel drums. When brief tones are produced by
nonpercussion instruments (e.g., staccato), there is a marked tendency to
increase the rate of successive tones. Contiguous staccato notes are seldom
separated by more than 1 s of silence. In such cases, echoic memory serves
to sustain the illusion of an uninterrupted line of sound.
As a general observation, we can note that in most of the worlds music
making there is a marked preference for some sort of continuous sound
activity; moreover, when instrumental sounds are brief in duration, such
sounds are often assigned to nonmelodic musical taskseven when the
tones produced evoke clear pitches.
3. Minimum Masking Principle
In his Nobel-prize-winning research, Georg von Bksy showed that different frequencies produce different points of maximum displacement on
the basilar membrane of the cochlea (Bksy, 1943/1949, 1960). Specifically, low frequencies cause the greatest displacement of the membrane
near the apex of the cochlea, whereas high frequencies produce maximum
displacements toward the oval window. Bksy, and later Skarstein
(Kringlebotn, Gundersen, Krokstad, & Skarstein, 1979), carefully mapped
this relationship by measuring the distance (in millimeters from the stapes)
of the point of maximum displacement for a given frequency input.5 This
correspondence between input frequency and place of maximum displacement of the basilar membrane is referred to as a tonotopic mapping or
cochlear map.
Subsequent work by Harvey Fletcher linked the frequency-place coordinates of the Bksy cochlear map to experimental data from frequency
discrimination and masking experiments (Fletcher, 1953, pp. 168175).
5. Greenwood (1961a) has produced the following function to express this frequencyposition relationship:
F = A(10ax - k),
where F is the frequency (in hertz), x is the position of maximum displacement (in millimeters from the apex), and A, a, and k are constants: in homo sapiens A = 165, a = 0.06, and
k = 1.0.
15
16
David Huron
Fig. 3. Approximate size of critical bands represented using musical notation. Successive
notes are separated by approximately one critical bandwidth = roughly 1-mm separation
along the basilar membrane. Notated pitches represent pure tones rather than complex
tones. Calculated according to the revised equivalent rectangular bandwidth (ERB) (Glasberg
& Moore, 1990).
tation (i.e., log frequency). Each note represents a pure tone; internote distances have been calculated according to the equivalent rectangular bandwidth rate (ERB) scale devised by Moore and Glasberg (1983; revised
Glasberg & Moore, 1990).
Plomp and Levelt (1965) extended Greenwoods work linking the perception of sensory dissonance (which they dubbed tonal consonance) to
the critical bandand hence to the mechanics of the basilar membrane.
Further work by Plomp and Steeneken (1968) replicated the dissonance/
roughness hypothesis using more contemporary perceptual data. Plomp
and Levelt estimated that pure tones produce maximum sensory dissonance
when they are separated by about 25% of a critical bandwidth. However,
their estimate was based on a critical bandwidth that is now considered to
be excessively large, especially below about 500 Hz. Greenwood (1991)
has estimated that maximum dissonance arises when pure tones are separated by about 30% to 40% of a critical bandwidth. For frequency separations greater than a critical band, no sensory dissonance arises between
two pure tones. These findings were replicated by Kameoka and Kuriyagawa
(1969a, 1969b) and by Iyer, Aarden, Hoglund, and Huron (1999).
In musical contexts, however, pure tones are almost never used; complex
tones containing several harmonic components predominate. For two complex tones, each consisting of (say) 10 significant harmonics, the overall
perceived sensory dissonance will depend on the aggregate interaction of
all 20 pure-tone components. The tonotopic theory of sensory dissonance
explains why the interval of a major third sounds smooth in the middle and
upper registers, but sounds gruff when played in the bass region.
Musicians tend to speak of the relative consonance or dissonance of
various interval sizes (such as the dissonance of a minor seventh or the
consonance of a major sixth). Although the pitch distance between two
complex tones affects the perceived sensory dissonance, the effect of interval size on dissonance is indirect. The spectral content of these tones, the
amplitudes of the component partials, and the distribution of partials with
respect to the critical bandwidth are the most formative factors determin-
17
ing sensory dissonance. Note that the characterizations of relative dissonance for fixed interval sizes given by musicians may also reflect cultural
(learned) factors that have received little overt empirical attention (Cazden,
1945).
Plomp and Levelt hypothesized that in the writing of chords, composers
would typically endeavor to maintain roughly equivalent amounts of sensory dissonance throughout the span of the chord. That is, they hypothesized that composers would typically avoid chords that produced disproportionately more dissonance in one or another region of the chord. A
given sonority might be highly dissonant or consonant overall, but Plomp
and Levelt supposed that the dissonance would typically be distributed
homogeneously across a given sonority.
An alternative interpretation of this prediction may be offered without
appealing to the concept of dissonance. Spectral components that lie within
a critical band of each other cause mutual masking, which reduces the
capacity of the auditory system to resolve or apprehend all of the sounds
present. If composers were disposed to reduce the capacity for mutual masking, then it would be appropriate to space chordal tones in such a way that
roughly equivalent amounts of spectral energy would fall in each critical
band. Since critical bandwidths span more semitones in the low region
than in the higher region, the notated pitches within a chord would show a
distinctive distribution were this hypothesis true.
Plomp and Levelt carried out a study of chordal-tone spacing in two
musical works: the third movement from J. S. Bachs Trio Sonata No. 2 for
organ (BWV 526)7 and the third movement from Dvorks String Quartet
Op. 51 in E major. Their analyses demonstrated an apparent consistency
between the composers arrangements of vertical sonorities and the tonotopic
mapthus implying that critical bands significantly influence the vertical
spacing of chordal tones.
In a replication study, Huron and Sellmer (1992) showed that Plomp
and Levelts demonstration was confounded by an unfortunate artifact and
that their results could not be used to support their conclusion. However,
using a more sophisticated inferential approach and a much larger musical
sample, we went on to provide an alternative demonstration that confirmed
Plomp and Levelts original hypothesis. The effect of spectral spread on the
spacing of chordal tones is not easily summarized; however, Figure 4 can
be used to illustrate the phenomenon. Figure 4 shows the average spacing
of notated (complex) tones for sonorities having various bass pitches from
C 4 to C2. For example, the first notated sonority in Figure 4 shows the
average tenor, alto, and soprano pitches for a large sample of four-notes
sonorities having C4 as the bass pitch. (The specific sonorities notated in
7. Incorrectly cited as Trio Sonata No. 3 by Plomp and Levelt (1965).
18
David Huron
Fig. 4. Average spacing of tones for sonorities having various bass pitches from C4 to C2.
Calculated from a large sample (>10,000) of four-note sonorities extracted from Haydn
string quartets and Bach keyboard works (Haydn and Bach samples equally weighted). Bass
pitches are fixed. For each bass pitch, the average tenor, alto, and soprano pitches are
plotted to the nearest semitone. (Readers should not be distracted by the specific sonorities
notated; only the approximate spacing of voices is of interest.) Note the wider spacing
between the lower voices for chords having a low mean tessitura. Notated pitches represent
complex tones rather than pure tones.
19
when the combined spectral content conforms to a single hypothetical harmonic series. This occurs most commonly when the frequencies of the component tones are related by simple integer ratios.
The pitch interval that most encourages tonal fusion is the aptly named
unison. The second most fused interval is the octave, whereas the third
most fused interval is the perfect fifth (DeWitt & Crowder, 1987; Stumpf,
1890). Following Stumpf, many music researchers have assumed that tonal
fusion and tonal consonance are the same phenomenon and that both arise
from simple integer frequency ratios. However, the extant psychoacoustic
research does not support Stumpfs view. Bregman (1990) has noted that
the confusion arises from conflating smooth sounding with sounding
as one. As we have seen, work by Greenwood (1961a, 1961b, 1990, 1991),
Plomp and Levelt (1965), Kameoka and Kuriyagawa (1969a, 1969b), and
Iyer et al. (1999) implicates critical band distances in the perception of
tonal consonance or sensory dissonance. This work shows that sensory
dissonance is only indirectly related to harmonicity or tonal fusion.
Whether or not tonal fusion is a musically desirable phenomenon depends on the music-perceptual goal. In Huron (1991b), it was shown that
in the polyphonic writing of J. S. Bach, tonally fused harmonic intervals are
avoided in proportion to the strength with which each interval promotes
tonal fusion. That is, unisons occur less frequently than octaves, which
occur less frequently than perfect fifths, which occur less frequently than
other intervals. Of course concurrent octaves and concurrent fifths occur
regularly in music, but (remarkably) they occur less frequently in polyphonic music than they would in a purely random juxtaposition of voices.
Note that this observation is independent of the avoidance of parallel
unisons, fifths, or octaves. As simple static harmonic intervals, these intervals are actively avoided in Bachs polyphonic works. Considering the importance of octaves and fifths in the formation of common chords, their
active avoidance is a remarkable feat (see Huron, 1991b).
In light of the research on tonal fusion, we may formulate the following
principle:
4. Tonal Fusion Principle. The perceptual independence of concurrent
tones is weakened when their pitch relations promote tonal fusion. Intervals that promote tonal fusion include (in decreasing order): unisons,
octaves, perfect fifths, ... Where the goal is the perceptual independence of concurrent sounds, intervals ought to be shunned in direct
proportion to the degree to which they promote tonal fusion.
20
David Huron
Fig. 5. Comparison of sensory consonance for complex tones (line) from Kaestner (1909)
with interval prevalence (bars) in the upper two voices of J.S. Bachs three-part Sinfonias
(BWVs 787801). Notice especially the discrepancies for P1 and P8. Reproduced from
Huron (1991b).
8. Technical justifications for sampling the upper two voices of Bachs three-part Sinfonias
are detailed by Huron (1994).
21
triads both contain one major third and one minor third each. However,
both the dominant seventh chord and the minor-minor-seventh chord contain one major third and two minor thirds. In addition, the diminished
triad (two minor thirds) is more commonly used than the augmented triad
(two major thirds). Hence, the increased prevalence of minor thirds compared with major thirds may simply reflect the interval content of common
chords. The relative prevalences of the major and minor sixth intervals
(inversions of thirds) are also consistent with this suggestion.
The discrepancies for P1 and P8 in Figure 5 are highly suggestive in light
of the tonal fusion principle. We might suppose that the reason Bach avoided
unisons and octaves is in order to prevent inadvertent tonal fusion of the
concurrent parts. In Huron (1991b), this hypothesis was tested by calculating a series of correlations. When perfect intervals are excluded from consideration, the correlation between Bachs interval preference and the sensory dissonance Z scores is 0.85.9 Conversely, if we exclude all intervals
apart from the perfect intervals, the correlation between Bachs interval
preference and the tonal fusion data is 0.82. Calculating the multiple regression for both factors, Huron (1991b) found an R2 of 0.88indicating
that nearly 90% of the variance in Bachs interval preference can be attributed to the twin compositional goals of the pursuit of tonal consonance
and the avoidance of tonal fusion. The multiple regression analysis also
suggested that Bach pursues both of these goals with approximately equal
resolve. Bach preferred intervals in inverse proportion to the degree to which
they promote sensory dissonance and in inverse proportion to the degree
to which they promote tonal fusion. It would appear that Bach was eager
to produce a sound that is smooth without the danger of sounding as
one. In Part III, we will consider possible aesthetic motivations for this
practice.
Musical Terminology: Types of Harmonic Intervals
The experimental results pertaining to sensory dissonance and tonal fusion may be used to illuminate traditional musical terminology. Music theorists traditionally distinguish three classes of harmonic intervals: perfect
consonances (such as perfect unisons, octaves, fourths, and fifths), imperfect consonances (such as major and minor thirds and sixths), and dissonances (such as major and minor seconds and sevenths, and tritones). These
interval types can be classified according to the criteria of sensory dissonance and tonal fusion. Perfect consonances typically exhibit low sensory
9. The word preference is not used casually here. Huron (1991b) developed an
autophase procedure that permits the contrasting of distributions in order to determine
what intervals are being actively sought out or actively avoided by the composer. The ensuing reported correlation values are all statistically significant.
22
David Huron
dissonance and high tonal fusion. Imperfect consonances have low sensory
dissonance and comparatively low tonal fusion. Dissonances exhibit high
sensory dissonance and low tonal fusion. (There are no equally tempered
intervals that exhibit high sensory dissonance and high tonal fusion, although the effect can be generated using grossly mistuned unisons, octaves,
or fifths.) It would appear that the twin phenomena of sensory dissonance
and tonal fusion provide a plausible account for both the traditional theoretical distinctions, as well as Bachs compositional practice.
5. Pitch Proximity Principle
In 1950, Miller and Heise observed that alternating pitches (such as trills)
produce two different perceptual effects depending on the pitch distance
separating the tones (Miller & Heise, 1950; see also Heise & Miller, 1951).
When the tones are close with respect to pitch, quick alternations evoke a
sort of undulating effectlike a single wavering line. However, when
the pitch separation is larger, the perceptual effect becomes one of two
beeping tones of static pitch. Musicians recognize this phenomenon as
that of pseudo-polyphony or compound melodic line, in which a single
sequence of pitches nevertheless evokes a sort of yodelling effect.
Miller and Heises observations were replicated and extended by a number of researchers including Bozzi and Vicario (1960) and Vicario (1960),
Schouten (1962), Norman (1967), Dowling (1967), van Noorden (1971a,
1971b), and Bregman and Campbell (1971). (Several of these researchers
worked independently, without knowledge of previously existing work.)
Of these pioneering efforts, the most significant works are those of Dowling
(1967) and van Noorden (1975). Over the past three decades, however, the
most sustained and significant research effort has been that of Albert
Bregman (1990).
In 1975 van Noorden mapped the relationship between tempo and pitch
separation on stream integration and segregation. Figure 6 summarizes van
Noordens experimental results. When the tempo is slow and/or the pitches
have close proximity, the resulting sequence is always perceived as a single
stream. This area is indicated in Figure 6 as region 1below the fission
boundary (lower line). Conversely, when the pitch distances are large and/
or the tempo is fast, two streams are always perceived. This condition is
indicated in Figure 6 as region 2to the left of the temporal coherence
boundary. Van Noorden also identified an intervening gray region, where
listeners may hear either one or two streams depending on the context and
the listeners disposition. Notice that the slope of the fission boundary is
much shallower than the slope of the temporal coherence boundary. We
will return to consider these different slopes later.
The importance of pitch proximity in stream organization is supported
by a wealth of further experimental evidence. Schouten (1962) observed
23
Fig. 6. Influence of interval size and tempo on stream fusion and segregation (van Noorden,
1975, p. 15). Upper curve: temporal coherence boundary. Lower curve: fission boundary. In
Region 1, the listener necessarily hears one stream (small interval sizes and slow tempos). In
Region 2, the listener necessarily hears two streams (large interval sizes and fast tempos).
24
David Huron
occurred when the transpositions removed all pitch overlap between the
concurrent melodies.
Deutsch (1975) and van Noorden (1975) found that, for tones having
identical timbres, concurrent ascending and descending tone sequences are
perceived to switch direction at the point where their trajectories cross.
That is, listeners are disposed to hear a bounced percept in preference to
the crossing of auditory streams. Figure 7 illustrates two possible perceptions of intersecting pitch trajectories. Although the crossed trajectories
represents a simpler Gestalt figure, the bounced perception is much more
commonat least when the trajectories are constructed using discrete pitch
categories (as in musical scales).
In summary, at least four empirical phenomena point to the importance
of pitch proximity in helping to determine the perceptual segregation of
auditory streams: (1) the fission of monophonic pitch sequences into pseudopolyphonic percepts described by Miller and Heise and others, (2) the discovery of information-processing degradations in cross-stream temporal
tasks, as found by Schouten, Norman, Bregman and Campbell, and
Fitzgibbons, Pollatsek, and Thomas, (3) the perceptual difficulty of tracking auditory streams that cross with respect to pitch described by Dowling,
Deutsch, and van Noorden, and (4) the preeminence of pitch proximity
over pitch trajectory in the continuation of auditory streams demonstrated
by Bregman et al. Stream segregation is thus strongly dependent upon the
proximity of successive pitches.
On the basis of extensive empirical evidence, we can note the following
principle:
5. Pitch Proximity Principle. The coherence of an auditory stream is
maintained by close pitch proximity in successive tones within the
stream. Pitch-based streaming is assured when pitch movement is within
van Noordens fission boundary (normally 2 semitones or less for
tones less than 700 ms in duration). When pitch distances are large, it
may be possible to maintain the perception of a single stream by reducing the tempo.
25
Fig. 8. Frequency of occurrence of melodic intervals in notated sources for folk and popular
melodies from 10 cultures (n = 181). African sample includes Pondo, Venda, Xhosa, and
Zulu works. Note that interval sizes correspond only roughly to equally tempered semitones.
26
David Huron
27
Fig. 9. Distribution of note durations in 52 instrumental and vocal works. Dotted line: note
durations for the combined upper and lower voices from J. S. Bachs two-part Inventions
(BWVs 772786). Dashed line: note durations in 38 songs (vocal lines only) by Stephen
Foster. Solid line: mean distribution for both samples (equally weighted). Note durations
were determined from notated scores using tempi measured from commercially available
sound recordings. Graphs are plotted using bin sizes of 100 ms, centered at the positions
plotted.
sample in order to avoid the effect of final lengthening and to make the
data consistent with the pitch alternation stimuli used to generate Figure 6.
A mean distribution is indicated by the solid line. In general, Figure 9
shows that the majority of musical tones are relatively short; only 3% of
tones are longer than 1 s in duration. Eighty-five percent of tones are shorter
than 500 ms; however, it is rare for tones to be less than 150 ms in duration. The median note durations for the samples plotted in Figure 9 are
0.22 s for instrumental and 0.38 s for vocal works.
This distribution suggests that the majority of tone durations in music
are too long to be in danger of crossing the temporal coherence boundarythat is, where two streams are necessarily perceived. Notice that compared with the temporal coherence boundary, the fission boundary is more
nearly horizontal (refer to Fig. 6). The boundary is almost flat for tone
durations up to about 400 or 500 ms after which the boundary shows a
positive slope. Figure 9 tells us that the majority of musical tones have
durations commensurate with the flat portion of the fission boundary. This
flat boundary indicates that intervals smaller than a certain size guarantee
the perception of a single line of sound. In the region of greatest musical
activity, the fission boundary approximates a constant pitch distance of
about 1 semitone.
28
David Huron
As in the case of harmonic intervals, music theorists traditionally distinguish two main classes of melodic intervals: conjunct or step motions, and
disjunct or leap motions. In Western music, the dividing line between step
and leap motion is traditionally placed between a major second and a minor third; that is, a major second (2 semitones) is considered a conjunct or
step motion, whereas a minor third (3 semitones) is considered the smallest
disjunct or leap motion. In other cultures, such as those that employ the
common pentatonic scale, the maximum step size is roughly 3 semitones.
It is plausible that the fission boundary in effect identifies a psychoacoustic
basis for the distinction between conjunct and disjunct melodic motions.
What theorists call conjunct intervals are virtually guaranteed to evoke the
perception of stream continuation.
Melodic MotionThe Temporal Coherence Boundary
Unlike the fission boundary, the temporal coherence boundary plotted
in Figure 6 shows a marked positive slope. This means that, in order for a
sequence of pitches to break apart into two streams, tempo is the predominant factoralthough pitch interval continues to play a significant role. If
the temporal coherence boundary influences musical organization, we might
expect to see tradeoffs between pitch interval and instantaneous tempo.
For example, we might predict that large pitch leaps would be associated
with tones of long duration.
Van Noorden (1975, p. 48) and Shepard (1981, p. 319) independently
drew attention to the similarity between Miller and Heises trill results and
Krtes third law of apparent motion in vision (Krte, 1915). Krte carried
out a number of experiments using two lamps that could be alternately
switched on and off. Krte found that the sense of apparent motion depends on the distance separating the two lamps and their speed of switching. If the lamps are placed farther apart, then the rate of switching must be
reduced in order to maintain a sense of apparent motion between the lamps.
If the switching rate is too fast, or the lamps are placed especially far apart,
then the viewer sees two independent flickering lights with no sense of
intervening motion. A possible explanation for this loss of apparent motion is that it is implausible for a single real-world object to move in accordance with the presumed trajectory.
The parallel between Krtes results and Miller and Heises trills is obvious and direct. It would seem that the sense of continuation between two
tones is an auditory analog to apparent motion in vision. Note that both
Krtes results and Miller and Heises results pertain to perception. Research by Huron and Mondor (1994) suggests that the cognitive basis for
this parallel can be attributed to the real-world generation or production
of motion.
29
In Huron and Mondor (1994), it was argued that both Krtes third law
of apparent motion and Miller and Heises results arise from the kinematic
principle known as Fitts law (Fitts, 1954). Fitts law applies to all musclepowered or autonomous movement. The law is best illustrated by an example (refer to Figure 10). Imagine that you are asked to alternate the
point of a stylus as rapidly as possible back and forth between two circular
targets. Fitts law states that the speed with which you are able to accomplish this task is proportional to the size of the targets and inversely proportional to the distance separating them. The faster speeds are achieved
when the targets are large and close togetheras in the lower pair of circles
in Figure 10.
Because Fitts law applies to all muscular motion, and because vocal
production and instrumental performance involve the use of muscles, Fitts
law also constrains the generation or production of sound. Imagine, for
example, that the circular targets in Figure 10 are arranged vertically rather
than horizontally. From the point of view of vocal production, the distance
separating the targets may be regarded as the pitch distance between two
tones. The size of the targets represents the pitch accuracy or intonation.
Fitts law tells us that, if the intonation remains fixed, then vocalists will be
unable to execute wide intervals as rapidly as for small intervals.
In Huron and Mondor (1994), both auditory streaming and melodic
practice were examined in light of Fitts law. In the first instance, two perceptual experiments showed that increasing the variability of pitches (the
auditory equivalent of increased target size) caused a reduction in the fission of auditory streams for rapid pitch alternationsas predicted by Fitts
Fig. 10. Two pairs of circular targets illustrating Fitts law. The subject is asked to alternate
the point of a stylus back-and-forth as rapidly as possible between two targets. The minimum duration of movement between targets depends on the distance separating the targets
as well as target size. Hence, it is possible to move more rapidly between the lower pair of
targets. Fitts law applies to all muscle motions, including the motions of the vocal muscles.
Musically, the distance separating the targets can be regarded as the pitch distance between
two tones, whereas the size of the targets represents pitch accuracy or intonation. Fitts law
predicts that if the intonation remains fixed, then vocalists will be unable to execute wide
intervals as rapidly as for small intervals.
30
David Huron
law. With regard to the perception of melody, Huron and Mondor noted
that empirical observations of preferred performance practice (Sundberg,
Askenfelt & Frydn, 1983) are consistent with Fitts law. When executing
wide pitch leaps, listeners prefer the antecedent tone of the leap to be extended in duration. In addition, a study of the melodic contours in a crosscultural sample of several thousand melodies showed them to be consistent
with Fitts law. Specifically, as the interval size increases, there is a marked
tendency for the antecedent and consequent pitches to be longer in notated
or performed duration. This phenomenon accords with van Noordens temporal coherence boundary. In short, when large pitch leaps are in danger of
evoking stream segregation, the instantaneous tempo for the leap is reduced. This relationship between pitch interval and interval duration is
readily apparent in common melodies such as My Bonnie Lies Over the
Ocean or Somewhere Over the Rainbow. Large pitch intervals tend to
be formed using tones of longer durationa phenomenon we might dub
leap lengthening.
It would appear that both Krtes third law of apparent motion and the
temporal coherence boundary in auditory streaming can be attributed to
the same cognitive heuristic: Fitts law. The brain is able to perceive apparent motion, only if the visual evidence is consistent with how motion occurs in the real world. Similarly, an auditory stream is most apt to be perceived when the pitch-time trajectory conforms to how real (mechanical or
physiological) sound sources behave. In short, it may be that Fitts law
provides the origin for the common musical metaphor of melodic motion (see also Gjerdingen, 1994).
Further evidence in support of a mechanical or physiological origin for
pitch proximity has been assembled by von Hippel (2000). An analysis of
melodies from cultures spanning four continents shows that melodic contours can be modeled very well as arising from two constraintsdubbed
ranginess and mobility. Such a model of melodic contour is able to account
for a number of melodic phenomena, notably the observation that large
leaps tend to be followed by a reversal in melodic direction.
6. Pitch Co-moduation Principle
As early as 1863, Helmholtz suggested that similar pitch motion contributes to the perceptual fusion of concurrently sounded tones. In recent
decades, Helmholtzs suggestion has received considerable empirical confirmation and extension. Chowning (1980) vividly demonstrated how coordinated frequency modulations would cause computer-generated voices
to fuse, whereas miscoordinated modulations would cause the sounds to
segregate. (This phenomenon was used in Chownings 1979 composition,
Phone.) Bregman and Doehring (1984) experimentally showed that tonal
31
fusion is enhanced significantly when two tones are modulated by correlated changes in log-frequency. Indeed, the degree of tonal fusion is greater
for tones of changing pitch than for tones of static pitch.
In an extensive series of experiments, McAdams (1982, 1984a, 1984b)
demonstrated that co-modulations of frequency that preserve the frequency
ratios of partials promote tonal fusion. Moreover, McAdams also showed
that positively correlated pitch motions that are not precise with respect to
log-frequency also tend to contribute to tonal fusion. In other words, tonal
fusion is most salient when co-modulation is precise with respect to logfrequency and the frequencies of the two tones are harmonically related.
Tonal fusion is next most salient when co-modulation is precise with respect to log-frequency and the frequencies of the two tones are not harmonically related. Finally, tonal fusion is next most salient when co-modulation is positively correlated, but not precise with respect to log-frequency.10
Apart from the empirical evidence, the pitch co-modulation principle is
evident in musical practice as well. In traditional music theory, theorists
distinguish two types of positively correlated pitch motion: similar motion
and parallel motion. These types of positively correlated pitch motions might
be collectively referred to as semblant motions. In Huron (1989a), it was
shown that polyphonic composers (not surprisingly) avoid semblant pitch
motionsboth parallel and similar contrapuntal motions. Moreover, it was
shown that parallel pitch motions are avoided more than similar motions.
Finally, it was shown that parallel motions are most avoided in the case of
intervals that tend most to promote tonal fusion: unisons, octaves, and
perfect fifths in particular. Once again, both traditional musical terminology and musical practice are consistent with the perceptual evidence.
Drawing on this research, we might formulate the following principle:
6. Pitch Co-modulation Principle. The perceptual union of concurrent
tones is encouraged when pitch motions are positively correlated. Perceptual fusion is most enhanced when the correlation is precise with
respect to log frequency.
32
David Huron
rive the voice-leading rules from the six principles described in the preceding section. The derivation given here is best characterized as heuristic rather
than formal. The purpose of this exercise is to clarify the logic, to avoid
glossing over details, to more easily expose inconsistencies, and to make it
easier to see unanticipated repercussions.
In the ensuing derivation, the following abbreviations are used:
G
A
C
D
goal
empirical axiom (i.e., an experimentally determined fact)
corollary
derived musical rule or heuristic
(traditional: commonly stated in music theory sources)
[D] derived musical rule or heuristic (nontraditional)
33
From this we derive one traditional, and one nontraditional rule of chordal
tone spacing:
D4. Chord Spacing Rule. In general, chordal tones should be spaced
with wider intervals between the lower voices.
[D5.] Tessitura-Sensitive Spacing Rule. It is more important to have
large intervals separating the lower voices in the case of sonorities that
are lower in overall pitch.
Recall that Huron and Sellmer (1992) showed that musical practice is indeed consistent with this latter ruleas predicted by Plomp and Levelt
(1965).
34
David Huron
The degree to which two concurrent tones fuse varies according to the
interval separating them.
A4a. Tonal fusion is greatest with the interval of a unison.
In general,
[D9.] Avoid Tonal Fusion Rule. Avoid unisons more than octaves, and
octaves more than perfect fifths, and perfect fifths more than other
intervals.
Therefore:
D10. Common Tone Rule. Pitch-classes common to successive sonorities are best retained as a single pitch that remains in the same voice.
A5b. The next best stream cohesion arises when pitch movement is near
or within van Noordens fission boundary (roughly 2 semitones or less).
35
Hence:
D11. Conjunct Movement Rule. If a voice cannot retain the same pitch,
it should preferably move by step.
C3. Avoid Leaps Rule. Avoid wide pitch leaps.
(This last rule is merely a corollary of the Common Tone and Conjunct
Movement rules.) When pitch movements exceed the fission boundary, effective voice-leading is best maintained when the temporal coherence boundary is not also exceeded:
A5c. Wide melodic leaps most threaten auditory stream cohesion when
they exceed the temporal coherence boundary.
Leap-lengthening (consistent with Fitts law) has been observed in a crosscultural sample of melodies by Huron and Mondor.
Bregman (1990) has characterized streaming as a competition between
possible alternative organizations. It is not simply a matter that two successive pitches need to be relatively close together in order to form a stream.
The pitches must be closer than other possible pitch-time traces. That is,
previous pitches compete for subsequent pitches:
A5d. Pitches are likely to stream to the nearest previous pitch.
From this, we can derive those rules of voice-leading that relate to pitchproximity competition:
D13. Nearest Chordal Tone Rule. Parts should connect to the nearest
chordal tone in the next sonority.
D14. Part-Crossing Rule. Avoid the crossing of parts with respect to
pitch.
36
David Huron
Once again, failure to obey this rule will conflict with the pitch proximity principle. Each pitch might be likened to a magnet, attracting the nearest subsequent pitch as its partner in the stream.11
Note that a simple general heuristic for maintaining within-voice pitch
proximity and avoiding proximity between voices is to place each voice or
part in a unique pitch region or tessitura. As long as voices or parts remain
in their own pitch territory, there is little possibility of part-crossing,
overlapping, or other proximity-related problems. There may be purely
idiomatic reasons to restrict an instrument or voice to a given range, but
there are also good perceptual reasons for such restrictionsprovided the
music-perceptual goal is to achieve optimum stream segregation. It is not
surprising that in traditional harmony, parts are normally referred to by
the names of their corresponding tessituras: soprano, alto, and so on.
Adding now the pitch co-modulation principle:
A6a. Effective stream segregation is thwarted when tones move in a positively correlated pitch direction.
[D16.] Semblant Motion Rule. Avoid similar or parallel pitch motion
between concurrent voices.
A6b. Effective stream segregation is especially thwarted when tones move
in precisely positively correlated pitch direction.
[D17.] Parallel Motion Rule. Avoid parallel motion more than similar
motion.
37
From this we derive the well-known prohibition against parallel fifths and
octaves:
D21. Parallel Unisons, Octaves, and Fifths Rule. Avoid parallel unisons,
octaves, or fifths.
Joining the tonal fusion, pitch proximity, and pitch co-modulation principles, we find that:
A4&A5b&A6. When moving in a positively correlated direction toward
a tonally fused interval, the detrimental effect of tonal fusion can be alleviated by ensuring proximate pitch motion.
From this we can derive the traditional injunction against exposed or hidden intervals.
38
David Huron
39
40
David Huron
Although musical notation provides only a crude indication of the temporal aspects of real performances, Raschs work suggests that synchronously and asynchronously notated events are likely to be perceived as
such. Of course tones that are notated as having concurrent onsets are by
no means performed precisely together. But in typical performances, such
tones are obviously more likely to promote the perception of tonal fusion
than tones that are notated as having asynchronous onsets. A difference of
a sixteenth-duration at a tempo of 80 quarter notes per minute produces a
nominal asynchrony of 187 ms, a time delay that is more than sufficient to
elicit the perception of separate events. In short, onsets that are deemed
asynchronous according to musical notation are unlikely to be perceived
as synchronous, whereas onsets that are deemed synchronous according
to musical notation may well be perceived as being synchronous. Thus
musical notation provides a useful (although not infallible) indication of
perceived onset synchronies and asynchronies in composed music.
Drawing on this research, we might formulate the following principle:
7. Onset Synchrony Principle. If a composer intends to write music in
which the parts have a high degree of perceptual independence, then
synchronous note onsets ought to be avoided. Onsets of nominally distinct sounds should be separated by 100 ms or more.
41
Fig. 11. Onset synchrony autophases for a random selection of 10 of Bachs 15 two-part
keyboard Inventions (BWVs 772786); see Huron (1993a). Values plotted at zero degrees
indicate the proportion of onset synchrony for the actual works. All other phase values
indicate the proportion of onset synchrony for rearranged musiccontrolling for duration,
rhythmic order, and meter. The dips at zero degrees are consistent with the hypothesis that
Bach avoids synchronous note onsets between the parts. Solid line plots a mean onset synchrony function for all data.
are more than two parts, the resulting autophases are multidimensional
rather than two dimensional. The four-part Adeste Fideles has been simplified to three dimensions by phase shifting only two of the three voices
with respect to a stationary soprano voice. As can be seen, in the Sinfonia
No. 1 there is a statistically significant dip at (0,0) consistent with the hypothesis that onset synchrony is actively avoided. In Adeste Fideles by
contrast, there is a significant peak at (0,0) (at the back of the graph) consistent with the hypothesis that onset synchrony is actively sought. The
graphs formally demonstrate that any realignment of the parts in Bachs
Sinfonia No. 1 would result in greater onset synchrony, whereas any realignment of the parts in Adeste Fideles would result in less onset synchrony.
Musicians recognize this difference as one of musical texture. The first
work is polyphonic in texture whereas the second work is homophonic in
texture. In a study comparing works of different texture, Huron (1989c)
examined 143 works by several composers that were classified a priori
according to musical texture. Discriminant analysis12 was then used in or12. Discriminant analysis is a statistical procedure that allows the prediction of group
membership based on functions derived from a series of prior classified cases. See Lachenbruch
(1975).
42
David Huron
der to discern what factors distinguish the various types of textures. Of six
factors studied, it was found that the most important discriminator between homophonic and polyphonic music is the degree of onset synchrony.
As we have noted, an auxiliary principle (such as the onset synchrony
principle) can provide a supplementary axiom to the core principles and
allow further voice-leading rules to be derived. Once again, space limitations prevent us from considering all of the interactions. However, consider the joining of the onset synchrony principle with the principal of tonal
fusion. In perceptual experiments, Vos (1995) has shown that there is an
interaction between onset synchrony and tonal fusion in the formation of
auditory streams.
43
A4&A7. When approaching a tonally fused interval, the detrimental effect of tonal fusion can be alleviated somewhat by ensuring asynchronous
tone onsets.
Fig. 13. Results of a comparative study of tonally fused intervals in polyphonic and (predominantly) homophonic repertoires by J. S. Bach. Only perfect harmonic intervals are
plotted (i.e., fourths, fifths, octaves, twelfths, etc.). As predicted, in polyphonic textures,
most perfect intervals are approached with asynchronous onsets (one note sounding before
the onset of the second note of the interval). In the more homophonic chorale repertoire,
most perfect intervals are formed synchronously; that is, both notes tend to begin sounding
at the same moment. By contrast, the two repertoires show no differences in their approach
to imperfect and dissonant intervals. From Huron (1997).
44
David Huron
Whence Homophony?
The existence of homophonic music raises an interesting perceptual
puzzle. On the one hand, most homophonic writing obeys all of the traditional voice-leading rules outlined in Part I of this article. (Indeed, most
four-part harmony is written within a homophonic texture.) This suggests
that homophonic voice-leading is motivated (at least in part) by the goal of
stream segregationthat is, the creation of perceptually independent voices.
However, because onset asynchrony also contributes to the segregation of
auditory streams, why would homophonic music not also follow this principle? In short, why does homophonic music exist? Why isnt all multipart
music polyphonic in texture?
The preference for synchronous onsets may originate in any of a number
of music-related goals, including social, historical, stylistic, aesthetic, formal, perceptual, emotional, pedagogical, or other goals that the composer
may be pursuing. If we wish to posit a perceptual reason to account for this
apparent anomaly, we must assume the existence of some additional perceptual goal that has a higher priority than the goal of stream segregation.
More precisely, this hypothetical perceptual goal must relate to some aspect of temporal organization in music. Two plausible goals come to mind.
In the first instance, the texts in vocal music are better understood when
all voices utter the same syllables concurrently. Accordingly, we might predict that vocal music is less likely to be polyphonic than purely instrumental music. Moreover, when vocal music is truly polyphonic, it may be better
to use melismatic writing and reserve syllable onsets for occasional moments of coincident onsets between several parts. In short, homophonic
music may have arisen from the concurrent pursuit of two perceptual goals:
the optimization of auditory streaming, and the preservation of the intelligibility of the text. Given these two goals, it would be appropriate to abandon the pursuit of onset asynchrony in deference to the second goal.
Another plausible account of the origin of homophony may relate to
rhythmic musical goals. Although virtually all polyphonic music is composed within a metric context, the rhythmic effect of polyphony is typically
that of an ongoing braid of rhythmic figures. The rhythmic strength associated with marches and dances (pavans, gigues, minuets, etc.) cannot be
achieved without some sort of rhythmic uniformity. Bachs three-part
Sinfonia No. 1 referred to in connection with Figure 12 appears to have
little of the rhythmic drive evident in Adeste Fideles. In short, homophonic music may have arisen from the concurrent goals of the optimization of
auditory streaming, and the goal of rhythmic uniformity. As in the case of
the text-setting hypothesis, this hypothesis generates a prediction
namely, that works (or musical forms) that are more rhythmic in character
(e.g., cakewalks, sarabandes, waltzes) will be less polyphonic than works
following less rhythmic forms.
45
Fig. 14. Voice-tracking errors while listening to polyphonic music. Solid columns: mean
estimation errors for textural density (no. of polyphonic voices); data for 130 trials from
five expert musician subjects. Shaded columns: unrecognized single-voice entries in polyphonic listening; data for 263 trials from five expert musician subjects (Huron, 1989b). The
data show that when listening to polyphonic textures that use relatively homogeneous timbres, tracking confusions are common when more than three voices are present.
46
David Huron
visual items exceeds three (Estes & Combes, 1966; Gelman & Tucker, 1975;
Schaeffer, Eggleston, & Scott, 1974). Descoeudres (1921) characterized this
as the un, deux, trois, beaucoup phenomenon. In a study of infant perceptions of numerosity, Strauss and Curtis (1981) found evidence of this phenomenon in infants less than a year old. Using a dishabituation paradigm,
Strauss and Curtis showed that infants are readily able to discriminate two
from three visual items. However, performance degrades when discriminating three from four items and reaches chance performance when discriminating four from five items. The work of Strauss and Curtis is especially significant because preverbal infants can be expected to possess no
explicit knowledge of counting. This implies that the perceptual confusion
arising for visual and auditory fields containing more than three items is a
low-level constraint that is not mediated by cognitive skills in counting.
These results are consistent with reports of musical experience offered
by musicians themselves. Mursell (1937) reported a conversation with Edwin
Stringham in which Stringham noted that no human being, however talented and well trained, can hear and clearly follow more than three simultaneous lines of polyphony. A similar claim was made by composer Paul
Hindemith (1944).
Drawing on this research, we might formulate the following principle:
8. Principle of Limited Density. If a composer intends to write music in
which independent parts are easily distinguished, then the number of
concurrent voices or parts ought to be kept to three or fewer.
47
Fig. 15. Relationship between nominal number of voices or parts and mean number of
estimated auditory streams in 108 polyphonic works by J. S. Bach. Mean auditory streams
for each work calculated using the stream-latency method described and tested in Huron
(1989a, chap. 14). Bars indicate data ranges; plotted points indicate mean values for each
repertoire. Dotted line indicates a relation of unity slope, where an N-voice polyphonic
work maintains an average of N auditory streams. The graph shows evidence consistent with
a preferred textural density of three auditory streams. See also Huron and Fantini (1989).
pseudo-polyphonic writing. Figure 15 shows that as the number of nominal voices increases, Bach gradually changes his compositional strategy.
For works with just two parts, Bach endeavors to keep the parts active (few
rests of short duration) and to boost the textural density through pseudopolyphonic writing. For works with four or more nominal voices, Bach
reverses this strategy. He avoids writing pseudo-polyphonic lines and retires voices from the texture for longer periods of time. Figure 15 reveals a
change of strategy consistent with a preferred textural density of three auditory streams. In light of the extant research, it would appear that Bach
tends to maximize the number of auditory streams, while simultaneously
endeavoring to avoid exceeding the listeners ability to track the concurrent
parts.
As for works that regularly use four or more concurrent parts, once
again, the notion of musical genre may be helpful in understanding diver-
48
David Huron
Fig. 16. Schematic illustration of Wessels illusion (Wessel, 1979). A sequence of three rising
pitches is constructed by using two contrasting timbres (marked with open and closed
noteheads). As the tempo of presentation increases, the percept of a rising pitch figure is
superceded by two independent streams of descending pitcheach stream distinguished by
a unique timbre.
49
Again, the question that immediately arises regarding this principle is,
why do composers routinely ignore it? Polyphonic vocal and keyboard
works, string quartets, and brass ensembles maintain remarkably homogeneous timbres. Of course, some types of music do seem to be consistent
with this principle. Perhaps the most common Western genre that employs
heterogeneous instrumentation is music for woodwind quintet. Many other
small ensembles make use of heterogeneous instrumentation. Nevertheless,
it is relatively rare for tonal composers to use different timbres for each of
the various voices. A number of reasons may account for this. In the case of
keyboard works, there are basic mechanical restrictions that make it difficult to assign a unique timbre to each voice. Exceptions occur in music for
dual-manual harpsichord and pipe organ. In the case of organ music, a
common Baroque genre is the trio sonata in which two treble voices are
assigned to independent manuals and a third bass voice is assigned to the
pedal division. In this case it is common to differentiate polyphonic voices
timbrally by using different registrationsindeed, this is the common performance practice. The two-manual harpsichord can also provide contrasting registrations, although only two voices can be distinguished in this
manner.
The case of vocal ensembles is more unusual. There are some differences
in vocal production, and hence there are modest differences in vocal character for bass, tenor, alto, and soprano singers. However, choral conductors normally do not aim for a clear difference in timbre between the various voices. On the contrary, most choral conductors attempt to shape the
sound of the chorus so that a more homogeneous or integrated timbre is
produced. Nevertheless, it is possible to imagine a musical culture in which
such vocal differences were highly exaggerated. Why isnt this commonplace?
We may entertain a variety of scenarios that might account for the avoidance of timbral differentiation in much music. Some of these scenarios relate to idiomatic constraints in performance. For example, much Western
polyphonic music is written for vocal ensemble or keyboard. On the keyboard, we have noted that simple mechanical problems make it difficult to
assign different timbres to each voice. In ensemble music, recruiting the
necessary assortment of instrumentalists might have raised practical difficulties. It might be easier to recruit two violinists than to recruit a violinist
and an oboe player. Such practical problems might have forced composers
50
David Huron
51
52
David Huron
AUXILIARY PRINCIPLES AND MUSICAL TEXTURES
In Part III of this article, we have argued that some perceptual principles
are auxiliary to the core voice-leading principles. It has been argued that
these principles are treated as compositional options and are chosen according to whether or not the composer wishes to pursue additional perceptual goals. If this hypothesis is correct, then measures derived from such
auxiliary principles ought to provide useful discriminators for various types
of music. For example, the degree of onset synchrony may be measured
according to a method described in Huron (1989a), and this measure ought
to discriminate between those musics that do, or do not, conform to the
principle of onset synchrony.
Figure 17 reproduces a simple two-dimensional texture space described
by Huron (1989c). A bounded two-dimensional space can be constructed
Fig. 17. Texture space reproduced from Huron (1989c). Works are characterized according
to two properties: onset synchronization (the degree to which notes or events sound at the
same time) and semblant motion (the degree to which concurrent pitch motions are positively correlated). Individual works are plotted as points; works within a single repertoire
are identified by closed figures.
53
Fig. 18. Distribution of musical textures from a sample of 173 works; generated using
Parzen windows method. (The axes have been rearranged relative to Figure 18 in order to
maintain visual clarity.)
54
David Huron
y points a Parzen windows technique (Duda & Hart, 1973) has been used
to estimate the actual distribution shapes. As can be seen, three classic
genresmonophony, polyphony, and homophonyare clearly evident.
(Heterophonic works were not sufficiently represented in the sample for
them to appear using this technique.)
By itself, this texture space does not prove the hypothesis that some
perceptual principles act to define various musical genres. This texture space
is merely an existence proof that, a priori, genres can be distinguished by 2
of the 10 principles discussed here. A complete explication of perceptually
based musical textures would require a large-scale multivariate analysis
that is beyond the scope of this article.
Conclusion
In this article, we have shown that perceptual principles can be used to
account for a number of aspects of musical organization, at least with respect to Western music. In the first instance, it was shown that six empirically established perceptual principles are able to account for the majority
of traditional rules in Western voice-leading. Moreover, the power of these
perceptual principles was evident in the fact that nontraditional rules derived from the principles were also found to be consistent with actual musical practice.
As in a formal system, it is possible to eliminate, modify, or add different
perceptual principlesin a manner analogous to axioms in formal logic.
The addition, elimination, or modification of different perceptual principles
leads to different sonic textures and so can be used to define or distinguish
various musical genres.
Of course, the account of voice-leading given here remains incomplete.
Unresolved issues include a variety of questions. First, why do theorists
typically restrict the injunction against exposed octaves to outer voices only?
Because inner voices are more difficult to trace perceptually, one might
have thought that inner voices would be more prone to stream integration
arising from exposed octaves. Second, little is known about the influence
of perceptual schemas in the formation of auditory streams. Bregman (1990)
has assembled evidence illustrating the existence and influence of learned
schemas in auditory scene analysis. It is possible that motivic and rhythmic
features of musical organization facilitate stream segregation, but no research has yet addressed this issue. For example, is it the case that imitative
polyphony facilitates or inhibits stream segregation compared with
nonimitative polyphony?
Third, the perceptual scenario outlined here does not account for the
sense of direction (leading) that attends musical pitch successions. For
55
56
David Huron
Of course, even if the rules of voice-leading were shown to be somehow immutable, their musical relevance depends entirely on the acceptance
of the underlying perceptual goals. In short, the composer must deem the
goals discussed here worth pursuing. If musicians see no need to pursue
these goals, then the principles described here are musically irrelevant. Even
when such goals are deemed worthy of pursuit, they must be balanced
against other possible compositional goalssuch as social, historical, stylistic, aesthetic, or formal goalsor other cognitive or perceptual goals
deemed to have a higher priority.
Aesthetic Origins
Given the plethora of possible compositional goals, we might rightly ask
why so much Western music appears to be consistent with the perceptual
goal discussed here. Why would so many composers endeavor to clarify
the auditory streaming of the individual parts in a musical texture?
Bregman (1990) has advanced the hypothesis that music makes use of
preexisting auditory mechanisms for parsing the environment. A useful
parallel can be drawn to animated cartoons. Although the cartoonist is free
to draw anything at all, there are advantages to drawing objects and motions that resemble real-world visual scenes. Viewers will readily apply preexisting visual mechanisms in making sense of fictitious images, even when
the objects of manipulation are nonrepresentational. According to Bregman,
composers similarly create auditory fictions populated by plausible (although largely fictional) auditory objects and events. Music listeners may
be predisposed to make use of their existing auditory expertise related to
the parsing of auditory scenes.
Absent from Bregmans account is an aesthetic discussion of why this
might be enjoyable for listeners. Why do many listeners take pleasure in
music constructed according to the rules of voice-leading?
Like all sensory systems, the auditory system arose as an evolutionary
adaptation whose function is to provide listeners with relevant information about the external world. Sensory systems are not infallible in carrying out this task, however. Sensory information can often be incomplete or
ambiguous, and deciphering the world entails a certain amount of mental
effort. Particularly when a scene is complicated, there is good reason to
suppose that the limbic system rewards the brain when perceptual systems
successfully parse the available information. By success, we do not
mean that the mental representation constructed by perception accurately
reflects the real worldsince there is no independent way of judging such
accuracy. Rather, by success we mean that the mental representation is
57
coherent, self-consistent, and congruent with information from other sensory systems and schematic expectations.
A number of phenomenal experiences accord with the view that pleasure is related to perceptual success. Watching an out-of-focus film can
engender considerable annoyance, followed by a sense of relief when the
focus is finally adjusted. A parallel auditory experience is evident when
tuning a radio to eliminate noise or static. Switching from monophonic to
stereophonic reproduction both improves the spatial resolution of the sound
sources and simultaneously evokes pleasure.
An apparent problem with this view arises from composers penchant to
create music containing several concurrent parts. If pleasure is evoked by
successful parsing of the auditory scene, then why wouldnt musicians create solely monophonic music? A plausible answer to this question is that
the amount of pleasure evoked may be proportional to the perceptual difficulty posed by the scene. Consider, by way of illustration, the experience
of random-dot stereograms. These are three-dimensional images that are
coded in a superficially confusing array of random dots. Seeing the threedimensional image can involve considerable mental effort, but if the image
is successfully apprehended, there ensues a rush of delight for the viewer.
Similarly, it is possible that some multipart music is organized to challenge
the listeners auditory parsing capacities.
This interpretation carries several implications. It suggests that early
Renaissance polyphonists discovered an aesthetic effect akin to randomdot stereograms: by challenging the listeners auditory parsing abilities, the
potential for a pleasing effect could be heightened. However, this heightened pleasure would be possible only if listeners could indeed successfully
parse the more complex auditory scenes. Having increased the perceptual
challenge, composers would need to take care in providing adequate streaming cues. Following the rules of voice-leading and limiting the density of
parts (as we have seen in Bach) might be essential aids for listeners. A
further implication is that at least some of the listeners who claim to dislike
polyphonic music may fail to get it in the same way that some viewers
fail to get random-dot stereograms.
This interpretation also has implications for how we understand sensory
dissonance. Earlier, we interpreted the congruence between chordal-tone
spacing and the cochlear map as consistent with minimizing masking. However, the results are equally consistent with the goal of minimizing sensory
dissonance. Since both sensory dissonance and auditory masking are linked
to the critical band, it may be that the phenomena are two sides of the same
coin. The central characteristic of dissonance is the negatively valenced
phenomenal experience of listenersakin to annoyance. Why would listeners find spectral components within a critical band annoying? Perhaps
58
David Huron
because the auditory system recognizes the presence of masking, with the
implication that any seemingly successful parsing of the auditory scene may
be deceptive or faulty.
Note that the preceding interpretation of the aesthetic origins of voiceleading should in no way be construed as evidence for the superiority of
polyphonic music over other types of music. In the first instance, different
genres might manifest different perceptual goals that evoke aesthetic pleasure in other ways. Nor should we expect pleasure to be limited only to
perceptual phenomena. As we have already emphasized, the construction
of a musical work may be influenced by innumerable goals, from social
and cultural goals to financial and idiomatic goals. Finally, it is evident
from Figure 17, that many areas of possible musical organization remain to be explored. The identification of perceptual mechanisms need not
hamstring musical creativity. On the contrary, it may be that the overt identification of the operative perceptual principles may spur creative endeavors to generate unprecedented musical genres.14,15
References
Aldwell, E., & Schachter, C. (1989). Harmony and Voice-Leading. 2nd ed. Orlando, FL:
Harcourt, Brace, Jovanovich.
ANSI. (1973). Psychoacoustical Terminology. New York: American National Standards
Institute.
Attneave, F., & Olson, R. K. (1971). Pitch as a medium: A new approach to psychophysical
scaling. American Journal of Psychology, 84, 147166.
von Bksy, G. (1943/1949). ber die Resonanzkurve und die Abklingzeit der verschiedenen
Stellen der Schneckentrennwand. Akustiche Zeitschrift, 8, 6676; translated as: On the
resonance curve and the decay period at various points on the cochlear partition. Journal of the Acoustical Society of America, 21, 245254.
von Bksy, G. (1960). Experiments in hearing. New York: McGraw-Hill.
Benjamin, T. (1986). Counterpoint in the style of J. S. Bach. New York: Schirmer Books.
Beranek, L. L. (1954). Acoustics. New York: McGraw-Hill. (Reprinted by the Acoustical
Society of America, 1986).
Berardi, A. (1681). Documenti armonici. Bologna, Italy: Giacomo Monti.
14. Several scholars have made invaluable comments on earlier drafts of this article.
Thanks to Bret Aarden, Jamshed Bharucha, Albert Bregman, Helen Brown, David Butler,
Pierre Divenyi (Divenyi, 1993), Paul von Hippel, Adrian Houstma, Roger Kendall, Mark
Lochstampfor, Steve McAdams, Eugene Narmour, Richard Parncutt, Jean-Claude Risset,
Peter Schubert, Mark Schmuckler, David Temperley, James Tenney, and David Wessel.
Special thanks to Gregory Sandell for providing spectral analyses of orchestral tones, to
Jasba Simpson for programming the Parzen windows method, and to Johan Sundberg for
drawing my attention to the work of Rssler (1952).
15. This article was originally submitted in 1993; a first round of revisions was undertaken
while the author was a Visiting Scholar at the Center for Computer Assisted Research in the
Humanities, Stanford University. Earlier versions of this article were presented at the Second
Conference on Science and Music, City University, London (Huron, 1987), at the Second
International Conference on Music Perception and Cognition, Los Angeles, February 1992,
and at the Acoustical Society of America 125th Conference, Ottawa (see Huron, 1993c).
59
60
David Huron
Dowling, W. J. (1967). Rhythmic fission and the perceptual organization of tone sequences.
Unpublished doctoral dissertation, Harvard University, Cambridge, MA.
Dowling, W. J. (1973). The perception of interleaved melodies. Cognitive Psychology, 5,
322337.
Duda, R. O., & Hart, P. E. (1973). Pattern classification and scene analysis. New York:
John Wiley & Sons.
Erickson, R. (1975). Sound structure in music. Berkeley: University of California Press.
Estes, B., & Combes, A. (1966). Perception of quantity. Journal of Genetic Psychology,
108, 333-336.
Ftis, F.-J. (1840). Esquisse de lhistoire de lharmonie. Paris: Revue et Gazette Musicale de
Paris. [English trans. by M. Arlin, Stuyvesant, NY: Pendragon Press, 1994.]
Fitts, P. M. (1954). The information capacity of the human motor system in controlling
amplitude of movement. Journal of Experimental Psychology, 47, 381391.
Fitzgibbons, P. J., Pollatsek, A., & Thomas, I. B. (1974). Detection of temporal gaps within
and between perceptual tonal groups. Perception & Psychophysics, 16(3), 522528.
Fletcher, H. (1940). Auditory patterns. Reviews of Modern Physics, 12, 4765.
Fletcher, H. (1953). Speech and hearing in communication. New York: Van Nostrand.
Fux, J. J. (1725). Gradus ad Parnassum. Vienna. [Trans. by A. Mann as The study of counterpoint. New York: Norton Co., 1971.]
Gauldin, R. (1985). A practical approach to sixteenth-century counterpoint. Englewood
Cliffs, NJ: Prentice-Hall.
Gelman, R., & Tucker, M. F. (1975). Further investigations of the young childs conception
of number. Child Development, 46, 167175.
Gjerdingen, R. O. (1994). Apparent motion in music? Music Perception, 11(4), 335370.
Glasberg, B. R., & Moore, B. C. J. (1990). Derivation of auditory filter shapes from notched
noise data. Hearing Research, 47, 103138.
Glucksberg, S., & Cowen, G. N., Jr. (1970). Memory for nonattended auditory material.
Cognitive Psychology, 1, 149156.
Greenwood, D. D. (1961a). Auditory masking and the critical band. Journal of the Acoustical Society of America, 33(4), 484502.
Greenwood, D. D. (1961b). Critical bandwidth and the frequency coordinates of the basilar membrane. Journal of the Acoustical Society of America, 33(4), 13441356.
Greenwood, D. D. (1990). A cochlear frequency-position function for several species29
years later. Journal of the Acoustical Society of America, 87(6), 25922605.
Greenwood, D. D. (1991). Critical bandwidth and consonance in relation to cochlear frequency-position coordinates. Hearing Research, 54(2), 164208.
Gregory, A. H. (1994). Timbre and auditory streaming. Music Perception, 12, 161174.
Guttman, N., & Julesz, B. (1963). Lower limits of auditory periodicity analysis. Journal of
the Acoustical Society of America, 35, 610.
Haas, H. (1951). ber den Einfluss eines Einfachechos auf die Hrsamkeit von Sprache.
Acustica, 1, 4958.
Hartmann, W. M., & Johnson, D. (1991). Stream segregation and peripheral channeling.
Music Perception, 9(2), 155183.
Heise, G. A., & Miller, G. A. (1951). An experimental study of auditory patterns. America
Journal of Psychology, 64, 6877.
Helmholtz, H. L. von (1863/1878). Die Lehre von der Tonempfindungen als physiologische
Grundlage fr die Theorie der Musik. Braunschweig: Verlag F. Vieweg & Sohn. [Translated as: On the sensations of tone as a physiological basis for the theory of music (1878);
second English edition, New York: Dover Publications, 1954.]
Hindemith, P. (1944). A concentrated course in traditional harmony (2 vols., 2nd ed.). New
York: Associated Music Publishers.
von Hippel, P. (2000). Redefining pitch proximity: Tessitura and mobility as constraints on
melodic interval size. Music Perception, 17(3), 315327.
von Hippel, P. & Huron, D. (2000). Why do skips precede reversals? The effect of tessitura
on melodic structure. Music Perception, 18(1), 5985.
61
Hirsh, I. J. (1959). Auditory perception of temporal order. Journal of the Acoustical Society
of America, 31(6), 759767.
Horwood, F. J. (1948). The basis of harmony. Toronto: Gordon V. Thompson.
Houtgast, T. (1971). Psychophysical evidence for lateral inhibition in hearing. Proceedings
of the 7th International Congress on Acoustics (Budapest), 3, 24 H15, 521524.
Houtgast, T. (1972). Psychophysical evidence for lateral inhibition in hearing. Journal of
the Acoustical Society of America, 51, 18851894.
Houtgast, T. (1973). Psychophysical experiments on tuning curves and two-tone inhibition. Acustica, 29, 168179.
Huron, D. (1987). Auditory stream segregation and voice independence in multi-voice musical
textures. Paper presented at the Second Conference on Science and Music, City University, London, UK.
Huron, D. (1989a). Voice segregation in selected polyphonic keyboard works by Johann
Sebastian Bach. Unpublished doctoral dissertation, University of Nottingham,
Nottingham, England.
Huron, D. (1989b). Voice denumerability in polyphonic music of homogeneous timbres.
Music Perception, 6(4), 361382.
Huron, D. (1989c). Characterizing musical textures. Proceedings of the 1989 International
Computer Music Conference (pp. 131134). San Francisco: Computer Music Association.
Huron, D. (1991a). The avoidance of part-crossing in polyphonic music: perceptual evidence and musical practice. Music Perception, 9(1), 93104.
Huron, D. (1991b). Tonal consonance versus tonal fusion in polyphonic sonorities. Music
Perception, 9(2), 135154.
Huron, D. (1991c). Review of the book Auditory scene analysis: The perceptual organization of sound]. Psychology of Music, 19(1), 7782.
Huron, D. (1993a). Note-onset asynchrony in J. S. Bachs two-part Inventions. Music Perception, 10(4), 435443.
Huron, D. (1993b). Chordal-tone doubling and the enhancement of key perception.
Psychomusicology, 12(1), 7383.
Huron, D. (1993c). A derivation of the rules of voice-leading from perceptual principles.
Journal of the Acoustical Society of America, 93(4), S2362.
Huron, D. (1994). Interval-class content in equally-tempered pitch-class sets: Common scales
exhibit optimum tonal consonance. Music Perception, 11(3), 289305.
Huron, D. (1997). Stream segregation in polyphonic music: Asynchronous preparation of
tonally fused intervals in J. S. Bach. Unpublished manuscript.
Huron, D., & Fantini, D. (1989). The avoidance of inner-voice entries: Perceptual evidence
and musical practice. Music Perception, 7(1), 4347.
Huron, D., & Mondor, T. (1994). Melodic line, melodic motion, and Fitts law. Unpublished manuscript.
Huron, D., & Parncutt, R. (1992). How middle is middle C? Terhardts virtual pitch
weight and the distribution of pitches in music. Unpublished manuscript.
Huron, D., & Parncutt, R. (1993). An improved model of tonality perception incorporating
pitch salience and echoic memory. Psychomusicology, 12(2), 154171.
Huron, D., & Sellmer, P. (1992). Critical bands and the spelling of vertical sonorities. Music
Perception, 10(2), 129149.
Iverson, P. (1992). Auditory stream segregation by musical timbre. Paper presented at the
123rd meeting of the Acoustical Society of America. Journal of the Acoustical Society of
America, 91(2), Pt. 2, 2333.
Iyer, I., Aarden, B., Hoglund, E., & Huron, D. (1999). Effect of intensity on sensory dissonance. Journal of the Acoustical Society of America, 106(4), 22082209.
Jeppesen, K. (1939/1963). Counterpoint: The polyphonic vocal style of the sixteenth century (G. Haydon. Transl.). Englewood Cliffs, NJ: Prentice-Hall, 1963. (Original work
published 1931)
Kaestner, G. (1909). Untersuchungen ber den Gefhlseindruck unanalysierter Zweiklnge.
Psychologische Studien, 4, 473504.
62
David Huron
Kameoka, A., & Kuriyagawa, M. (1969a). Consonance theory. Part I: Consonance of dyads. Journal of the Acoustical Society of America, 45, 14511459.
Kameoka, A., & Kuriyagawa, M. (1969b). Consonance theory. Part II: Consonance of
complex tones and its calculation method. Journal of the Acoustical Society of America,
45, 14601469.
Keys, I. (1961). The texture of music: From Purcell to Brahms. London: Dennis Dobson.
Krte, A. (1915). Kinomatoscopishe Untersuchungen. Zeitschrift fr Psychologie der
Sinnesorgane, 72, 193296.
Kringlebotn, M., Gundersen, T., Krokstad, A., & Skarstein, O. (1979). Noise-induced hearing losses. Can they be explained by basilar membrane movement? Acta OtoLaryngologica Supplement, 360, 98101.
Krumhansl, C. (1979). The psychological representation of musical pitch in a tonal context.
Cognitive Psychology, 11, 346374.
Kubovy, M., & Howard, F. P. (1976). Persistence of pitch-segregating echoic memory. Journal of Experimental Psychology: Human Perception & Performance, 2(4), 531537.
Lachenbruch, P. A. (1975). Discriminant Analysis. New York: Hafner Press.
Martin, D. (1965). The evolution of barbershop harmony. Music Journal Annual, p. 40.
Mayer, M. (1894). Researches in Acoustics. No. IX. Philosophical Magazine, 37, 259288.
McAdams, S. (1982). Spectral fusion and the creation of auditory images. In M. Clynes
(Ed.), Music, mind, and brain: The neuropsychology of music (pp. 279298). New York
& London: Plenum Press.
McAdams, S. (1984a). Segregation of concurrent sounds. I: Effects of frequency modulation coherence. Journal of the Acoustical Society of America, 86, 21482159.
McAdams, S. (1984b). Spectral fusion, spectral parsing and the formation of auditory images. Unpublished doctoral dissertation, Stanford University, Palo Alto, CA.
McAdams, S., & Bregman, A. S. (1979). Hearing musical streams. Computer Music Journal, 3(4), 2643, 60, 63.
Merriam, A. P., Whinery, S., & Fred, B. G. (1956). Songs of a Rada community in Trinidad.
Anthropos, 51, 157174.
Miller, G. A., & Heise, G. A. (1950). The trill threshold. Journal of the Acoustical Society of
America, 22(5), 637638.
Moore, B. C. J., & Glasberg, B. R. (1983). Suggested formulae for calculating auditoryfilter bandwidths and excitation patterns. Journal of the Acoustical Society of America,
74(3), 750753.
Morris, R. O. (1946). The Oxford harmony. Vol. 1. London: Oxford University Press.
Mursell, J. (1937). The psychology of music. New York: Norton.
Narmour, E. (1991). The Analysis and cognition of basic melodic structures: The implication-realization model. Chicago: University of Chicago Press.
Narmour, E. (1992). The analysis and cognition of melodic complexity: The implicationrealization model. Chicago: University of Chicago Press.
Neisser, U. (1967). Cognitive psychology. New York: Meredith.
van Noorden, L. P. A. S. (1971a). Rhythmic fission as a function of tone rate. IPO Annual
Progress Report, 6, 912.
van Noorden, L. P. A. S. (1971b). Discrimination of time intervals bounded by tones of
different frequencies. IPO Annual Progress Report, 6, 1215.
van Noorden, L. P. A. S. (1975). Temporal coherence in the perception of tone sequences.
Doctoral dissertation, Technisch Hogeschool Eindhoven; published Eindhoven: Druk
vam Voorschoten.
Norman, D. (1967). Temporal confusions and limited capacity processors. Acta Psychologica,
27, 293297.
Ohgushi, K., & Hato, T. (1989). On the perception of the musical pitch of high frequency
tones. Proceedings of the 13th International Congress on Acoustics (Belgrade), 3, 27
30.
Opolko, F., & Wapnick, J. (1989). McGill University Master Samples. Compact discs Vols.
111. Montreal: McGill University, Faculty of Music.
63
64
David Huron