Date post: | 24-Jan-2023 |
Category: |
Documents |
Upload: | canterbury |
View: | 0 times |
Download: | 0 times |
1
The Musical Brain: a Study
into Musical Perception
and Memory
By Martha Nye
Canterbury Christ Church University
Department of Music and Performing Arts
Bachelor of Music
Word Count: 11, 552
Date of Submission: Friday 3rd
May 2013
2
Abstract This paper explores the neurological basis for the perception and memory for pitch, pitch
relationships and melodies. The phenomenon that musicians with similar training and
expertise excel in different areas of music leads to a question: if the training is similar, then
how is trainee or his methods different? This research explores the difference between, and
reasons for aptitude in, aural recall and sight-reading. The study shows the brain areas and
processes specific to music and the flexibility of musicians’ methods. The findings lead to a
personal understanding of proficiency or ineptitude in these skills.
3
Table of Contents
Introduction ...................................................................................................................................... 4
Musical Neurophysiology ..................................................................................................................... 5
Chapter 1 – Pitch and Melody ..................................................................................................... 8
Recognition of Pitch ............................................................................................................................. 8
Memory Systems for Pitch ................................................................................................................. 11
Absolute Pitch ..................................................................................................................................... 14
Melody and Pitch Recognition ........................................................................................................... 16
Immediate Recall of Melodies ............................................................................................................ 18
Chapter Summary ............................................................................................................................... 21
Chapter 2 – Melody and Harmony ........................................................................................... 23
Melodic Contour ................................................................................................................................. 23
Scale Structure and Tonality ............................................................................................................... 24
Musical Notation and Structure .......................................................................................................... 27
Chapter Summary ............................................................................................................................... 34
Chapter 3 – Music in Action ....................................................................................................... 36
Sight-Reading and Working Memory ................................................................................................. 36
Imagery and Mental Rehearsal ........................................................................................................... 37
Brain Function and Structure .............................................................................................................. 40
Chapter Summary ............................................................................................................................... 40
Conclusion ...................................................................................................................................... 42
References ....................................................................................................................................... 44
4
Introduction From the age of 18, my interest in brain organisation, function and mechanisms has grown,
especially in the field of music. Music has an emotional effect, an automatic effect and often
an unnoticed effect on our thoughts and feelings. The fact that music seems a phenomenon,
separate and far less understood from other brain studies, has appealed to me and led me to
search into the anatomical relationships of musical activity in the brain. My first interest was
that of ability – why could I easily pick up instruments, the reading of musical notation and
the act of singing when my peers, exposed to a similar level of musical activity, could not?
Why can I, after a similar amount and intensity of training to my sister, sight-read to near
perfection when she cannot? Yet she excels in aural tests, such as immediate recall of
melody and rhythm, and I perform poorly. What is it about our brains and their structure and
organisation that leads us to excel in different aspects of music? Which parts of my brain
have developed further or lesser to my sibling’s?
The study of memory has long interested me, both inside and outside of music, and
memory is intrinsically linked to the method and efficacy of perception and coding. A
particular issue of mine is that I cannot, and never have been able to, sing or play back a
melody I have just heard. As a music student I daily engage in many musical activities,
often spending a large proportion of my day exercising those musical structures in my brain.
I wanted to know which parts of my brain are stopping me either properly encoding or
remembering pitches and pitch sequences. The study of perception of melody may explain
to me where the deficiencies are. Perhaps the part of my brain which is lacking is also
implicated in other non-musical functions which are poor in performance, and this study
will lead me to understand where I excel and where I do not, and why.
5
Musical Neurophysiology Music neurophysiology is an excellent tool for exploring brain structure, function and
organisation. Despite the fact that many become involved in musical training, only a small
percentage of the population develop a proficiency in music, which suggests that there is an
innate predisposition to musical skill or that something in a musician's brain plasticity is
different in nature to a non-musician’s (Peretz and Zatorre 2005). Similarly, the study of
brain organisation is a useful aid to the study of how the brain functions in music. Studies of
brain anomalies show the use of neural networks for music that are separated from non-
musical auditory and vocal functions. Using these two together we can discover some
neurological roots to a very human trait (Peretz and Zatorre 2005).
Many different parts of the brain are found to have been active during activities of
music processing. Just from the first sound entering the ear, structures like the cochlea,
brain stem, mid brain nuclei and mid brain cortex are activated and in a very short space of
time lead to perception. Different parts of the brain are involved in stages of the musical
processing sequence, and there has been evidence of segregation between these mechanisms
(Peretz and Zatorre 2005). The organisation of the systems within the cochlea (See Figures
1 and 2) appears to correspond to frequencies. The basal parts respond to high frequencies
while the apical parts respond to lower frequencies (Weinberger 1999). Music is processed
like any other auditory information, through the ascending auditory pathway to the auditory
cortex. The cortex is on the superior temporal plane within the Sylvian fissure, demonstrated
by the thick line on Figure 3. The primary auditory cortex is in the superior temporal gyrus,
numbered 41 and 42 on the diagram (Figure 3). Laterally positioned to the primary auditory
cortex are the secondary auditory cortical areas (Stewart et al 2006).
6
In order to recognise and analyse musical information entering the brain the musical
working memory system must produce an internal representation. This trace must be
independent of the ‘superficial variations’ like dynamics and reverberation because music
relies on the relationships between the pitch values and not the independent pitch values. In
Figures 1 and 2:
Powerhousemuseum.com
(2001) and Skidmore.edu
(n.d.). Inner ear structure
showing the position of
the cochlea and its relation
to brain structures.
7
other words, it is recognition of the relationships between the pitches that make a melody
familiar (Peretz and Zatorre 2005).
Figure 3: a diagram of the brain areas.
8
Chapter 1 – Pitch and Melody
Recognition of Pitch Pitch is one the fundamental elements of music, along with rhythm, and yet their processing
functions appear to be independent, whilst still relying on each other to make a melody.
Jones and Boltz found that in fully functioning brains, perception and memory of pitch is
quite rhythmical - perhaps aided by the rhythm – but that brain damage can cause a
deficiency in pitch comprehension and perception while the metric and rhythmic skills
remain intact (Peretz and Zatorre 2005, 91). This has been shown in a case study by
Fasanaro et al (1990). The subject was a 72-year-old professional musician playing several
instruments and was experienced in sight-reading. He suffered a loss of strength to his right
side and space-time disorientation. Neurological examinations showed mild right
hemiparesis, a weakness on one side of the body, including reduced sense of touch, sight
and a difficulty reading and naming. A brain scan showed decreased density in the
temporoparieto-occipital region (the junction between temporal, parietal and occipital
areas), with damage in the left medial region and splenium (in the very centre of the brain).
The following table (Table 1) shows how he performed in the musical ability tests and how
his pitch abilities are far decreased in comparison to his rhythmical abilities.
9
A logical suggestion to make following this evidence is that pitch and rhythm
dimensions use separable subsystems within the musical mechanisms. Griffiths, Johnsrude,
Dean and Green (1999, in Peretz and Zatorre 2005) discovered similar activation patterns in
the brain in tests of pitch and time discrimination. Recall is never note perfect while it can
be harmonically and metrically consistent (Sloboda and Parker 2005). Evidence suggesting
that the operations are separable suggests that the brain areas that they do share are capable
of several simultaneous functions.
In studying how we compute pitch we can see that the right temporal neo-cortex,
located in the transverse temporal gyri, is heavily implicated. It is easy to see the function of
brain areas when comparing brain-damaged subjects to undamaged subjects. In a study by
Milner in 1962, those with focal brain excisions in the right temporal area were found to
have worse pitch relationship skills than those with excisions in the left. The impairment
was more than just a deficit in pitch discrimination, suggesting that it is implicated in more
than just this function (Peretz and Zatorre 2005, 91).
Table 1: Fasanaro
et al (1990, 268)
case study. A
table presenting
the results of the
subject’s musical
ability tests.
10
In Heschl's gyri, (the area containing the temporal gyri in Figure 3), damage in the
right anterolateral section can cause further damage to pitch processing. Perception of notes
missing their fundamental is difficult, but subjects with this damage seem to be better at
determining the direction of pitch (Peretz and Zatorre, 2005). For recognition of a
frequency, the fundamental tone is not crucial as it can be identified without. For lower
notes, the high frequencies are more important and for high notes, the lower frequencies are
more dominant (Pierce 1999).
The posterior secondary auditory cortex may be involved in the analysis of pitch, as
it was shown to be active when comparing complex tones (a tone with harmonics), which
include some spectral change. Manipulation of this pitch leads to more response in the right
auditory regions, with manipulation of pitch height activating the posterior section and
analysis of pitch chroma (its qualities) taking place in the anterior section (Peretz and
Zatorre 2005, 92). All of this evidence points to the right secondary auditory cortex’s role in
processing of pitch relationships. This occurs even when attention is directed elsewhere as
Tervaniemi (2003) showed with electroencephalography (EEG) and magneto-
encephalography (MEG) scans, where activity in the brain was still occurring, although to a
lesser extent. In Tervaniemi’s studies, she found that ‘temporally and spectrally complex
sounds as well as their relations are automatically represented in the human auditory cortex’
(Tervaniemi 2003, 294) and that musical sounds are distinct from phonetic sounds in their
neural representations. The brain has an ability to encode and analyse complex sounds even
when the subject is focused on another task, and musical sounds continue to be differently
analysed (Tervaniemi 2003).
Pitch relationships include the intervals between them, and there are several means
of pitch and interval identification. ‘Magnitude Estimation’ occurs when the listeners group
close frequencies into the same magnitude (pitch) and then jump to another magnitude when
11
the pitch gets too far out of range. ‘Absolute Identification’ is a skill held by those with
absolute pitch. These people can identify approximately 75 items, compared with those who
do not possess absolute pitch who can identify approximately 5 items. In interval
recognition for those with absolute pitch, ascending and descending intervals may be
perceived as the same, as it is merely the pitch categories that are needed for recognition.
The musicians tested could recognise the intervals (even those only with relative pitch) from
unison to a major tenth with 100% accuracy. ‘Category scaling’ meant the participants
shifted slightly narrower or wider intervals into equal temperament (Burns 1999, 221).
Memory Systems for Pitch Memory for pitch is a crucial part to our perception, comprehension and analysis of pitch
and melody. Whilst the distinct scale steps are not coded in the memory trace, they must be
perceived in order to make the representative trace, built on a framework of tonality and
contour (Dowling 1978). Music unfolds and develops over a time period, and therefore our
auditory mechanisms must maintain information correctly in order to relate certain elements
to those previous and subsequent. Deutsch (1970) found in her studies that the systems for
maintenance of pitch are dissociable from other memory parts, similarly to its perception as
we have seen above. A melody is analysed in terms of its pitch, dynamics, duration,
harmonic and melodic intervals, durational relationships, timbres, metre and other musical
elements by subdivisions, and each subdivision contains information on pitch, duration or
other elements. These are stored in parallel and the outputs combine at the time of retrieval
(Deutsch 1970). When one is determining whether two pitches are the same, the task
remains easy after a short interval. Although decay of the memory trace may have started, it
is not significant. When other tones are interpolated within a five-second delay, the task
becomes very difficult. The more tones that are interpolated, the more difficult the task
becomes. Pitch memory here is subject not only to time decay, but to interference. If
12
difficulty arose from distraction of attention, then interpolated material of all kinds would
result in memory loss, but this did not happen. If it were held in a store with limited
capacity, then other non-musical materials would also be forgotten in the overload, but this
did not happen. A specialised pitch system is becoming overloaded with pitch information
only (Deutsch 1970).
This theory was tested in four groups. They were presented with a test pitch, a five-
second interval, and then a comparison pitch, which they then had to compare to the test
pitch. In group 1) six tones were interpolated in the interval before pitch comparison; in
group 2) six numbers were spoken in the interval before pitch comparison; in group 3) they
had the same as number 2, but were requested to both compare the pitches and recall the
numbers spoken in the interval; and in group 4) they were required to recall the numbers and
not compare the pitches. Table 2 shows the results.
Since group 3 had to memorise both numbers and pitches and had comparable pitch
memory impairment to group 2, and recalled numbers equally well as group 4, this shows
the different information was likely coded in separate systems. Group 1 had pitch memory
impairment from interpolated pitch, but no group had pitch memory impairment from
interpolated speech. Impairment could not be due to the lack of rehearsal as when speech
was interpolated error was far smaller. This suggests that pitch was processed in a
Table 2: Percentage error in
groups 1 to 4 (Deutsch 1970).
Group 1 shows substantial
impairment on pitch memory,
groups 2 and 3 show minimal
pitch memory impairment and
groups 3 and 4 show comparable
number recall error.
13
specialised system, which was subject to interference from other pitch material, but not non-
musical material such as the spoken numbers (Deutsch 1970).
Deutsch (1973) also tested the error rate in pitch comparison when interpolated tones
were displaced by an octave. The tones of the same octave caused the greatest interference,
but the tones of the octave above were also highly disruptive. The least interference came
from the octave below. This could be explained by a suggestion that higher pitches attract
more attention, or that there is an independent process for interval analysis that is separate to
that of octave generalisation (Deutsch 1972) (and as we know this happens automatically
(Tervaniemi 2003)), meaning it becomes difficult to manage both tasks. The highest error
came from the test by Deutsch in 1974 when interpolated tones were played from all three
octaves, as the interval size was the largest (Deutsch 1999, 400).
Zatorre et al (1994) found that several regions are activated when working memory
load is high and that the frontal cortical region and posterior temporal regions are activated
in working memory for tones. Through measurements of Cerebral Blood Flow (CBF)
Zatorre and his team concluded that specialised neural systems that dealt with perceptual
analysis of melody were placed in the right superior temporal gyrus and pitch comparison is
computed in a specialised system in the right prefrontal cortex while retention of pitch
involves the right temporal and frontal cortices.
The auditory cortical areas are involved in high-information-load rehearsal, such as
in sight-reading, where one needs to retain pitch information while new information is
presented. In a study by Gaab et al (2003), subjects listened to six tones and were required
to judge whether the first and last tones were the same or different. Analysis of the
comparison caused activation patterns in the superior temporal gyrus and left inferior frontal
14
gyrus, among other regions, measured through functional Magnetic Resonance Imaging
(fMRI) scans. These studies point to the existence of a working memory pitch subsystem.
Absolute Pitch A working memory pitch subsystem would involve the analysis of single pitches and related
pitches using the skills Relative Pitch (RP) and Absolute Pitch (AP). RP is the ability to
calculate pitches from their relation with a known pitch. The ability to recognise a pitch
without reference to another is known as perfect or absolute pitch (Bermudez and Zatorre
2009). Those who possess this trait report that it was not a learnt ability and that it was
acquired incidentally, which suggests they have a predilection affecting their perception of
information. It is likely that these ‘cognitive inclinations’ (Bermudez and Zatorre 2009) are
a part of the developmental, tuitional and genetic influences that lead to the development of
AP. Predispositions are very likely to be involved. Baharloo et al (2000) found sibling
recurrence of AP to be quite high. This suggests an innate neural substrate interacts with the
environment at a certain time in development. All possessors have been engaged in early
musical training (from under the age of six). This is when the central nervous system is most
adaptable and there seems to be more effect on the structures associated with the musical
activity at this age (Altenmueller and Schneider 2009). In their study of 113 siblings of AP
possessors, 50% also had the youthful training, and 50% of those also reported possessing
AP.
Miyazaki (1992) reported that AP impeded the development of RP, which is a more
important skill in musically meaningful situations. AP possessors were slower to recognise
or calculate intervals when the starting note was ‘out of tune’, suggesting that they were
perceived as two separate tones rather than a tonal relation. Several of the participants
counted up from one recognised note to the next in order to calculate the interval. However,
15
many musicians have both AP and RP. AP was seen to be more prevalent in those deemed
to be extremely talented musicians and performers.
The brain activity of those possessing AP can be contrasted with those who do not,
as there is reduced electrical brain activity used in the working memory tasks of those
possessing AP. There also seems to be distinctly less response to pitch change. The right
frontal cortex, involved in monitoring pitch information in the working memory, is also less
active in those with AP. This suggests an element of efficiency or a lack of sensitivity, as
there is less analysis of pitch to be done (Peretz and Zatorre 2005). This is likely linked with
the categorical representation of pitch associated with AP as opposed to the continuous
trace, which allows for RP recognition. The right superior temporal gyrus and right inferior
frontal regions are activated when music is presented, and these are implicated in the
analysis of tonality. A link to tonal processing suggests that the initial perception of pitch
and auditory information is no different for a possessor of AP. The difference came when
single pitches were presented and activity in the left posterior dorsolateral frontal cortex was
noted in AP groups but not RP groups. This portion of the brain is associated with
conditional associative learning - learnt associations between a task and the appropriate
response (Sacchietti et al 2008; Zatorre et al 1998) - which may lead to a definition of AP as
the link of a stimulus (the pitch) to a response (the label or pitch name). This area was also
active in both groups when judging major or minor chords, which leads the authors to a
hypothesis that this area is implicated in verbal-tonal associations (Zatorre et al 1998).
The evidence of functional differences in the brain suggests that there may also be
some structural differences. A lateral asymmetry of the auditory cortical areas, which has
been evident in many studies, seems to be present in those possessing AP. There appears to
be either an enhancement of the left hand side or conversely a reduction on the right hand
side, which causes the difference (Keenan et al 2001). It is speculated that the early musical
16
training is an influence of structural development. The size of the right-hand side was far
more consistent in the AP musicians tested than the left compared to the non musicians and
the RP musicians, suggesting that it is a ‘pruning’ of the right hand side which is the
influence on the development of AP (Keenan et al 2001, 1407).
Melody and Pitch Recognition In order to recognise melodies, a long-term trace of the pitch/time relations must be made
and maintained. Music is not based on a semantic system like language and instead the
representation must be one of form and structure. Melodies can be recognised at any pitch
and so are remembered as schema, not the absolute pitches, involving the intervals, contour
and scale degree context (Snyder 2009). We cannot use meaning to remember the melody –
meaning is conveyed through other means, such as our associations with memories and
emotions. As melodies are recognisable in different transpositions and with tempo and
instrumental change, the traces are likely to be abstract (Peretz and Zatorre 2005).
Processing of the semantics of music takes place in the middle temporal gyrus and the left
anterior temporal area (Koelsch and Siebel 2005).
However, Halpern discovered that we do often preserve some of the surface features
of music such as dynamics, tempo and some retention of the absolute pitch of tunes despite
insufficient verbal or visual cues for the pitch (Halpern 1989). Our memory’s duality in this
respect parallels our perception of music. We will never process every detail but we must be
aware of and consider the surface details of each unique performance of a piece. Pitch and
precise tempo can be retained in some form, but the memory of the entire piece is
remembered as more of a ‘gist’. The trace follows the structure of organisation (Peretz and
Zatorre 2005, 97) and uses the surface elements in order to relate these sections (Lamont
and Dibben 2001), which are used depending on the piece’s most memorable elements.
Lamont and Dibben (2001) also found that listeners are less likely to use more complex
17
musical points, such as motivic development, in order to create mental representations of the
work. However, in order to appreciate and understand a unique performance, the surface
details like tempo and absolute pitch are required to be analysed. Both the structural and
surface details can be retained.
Further support for a musical memory subsystem comes from lesion studies. The
perception can remain intact whilst recognition of familiar melodies becomes impossible.
Sometimes this is even limited to music. Peretz (1996, 484) studied one lady who had a
bilateral auditory cortex lesion and was considered amusical (an inability to comprehend or
respond to music). She had a normal memory for lyrics when spoken, but was unable to
recognise the corresponding melodies. She had no other non-musical auditory deficits or
visual deficits, suggesting the damage was limited to this musical function. Zatorre and
Samson in 1992 found that after a right-sided temporal lobe lesion, retention of musical
memory quickly deteriorates. However, this right side is not as critical when the melody is
highly familiar. When recognition of a novel melody becomes difficult after damage to the
right temporal lobe, the left temporal lobe is used (Peretz and Zatorre 2005).
Bower in 1991 believed that mood might also have an effect on our memory for
music. Participants were placed in a positive and negative mood and then listened to a
melody judged as being happy or sad. In recognition there was a mood congruency effect
whereby the sad melodies were remembered better by those who were placed in a negative
mood, and vice versa (Houston and Haddock 2007). Houston and Haddock also found that
mood congruency effects are present in verbal and visual memory. This relates to the
activity on the right and left side of the amygdala that affects our moods, and this is held
deep inside the brain and connected to the auditory cortex (Scott et al 2000).
18
Immediate Recall of Melodies Aural tests in instrumental exams require candidates to sing back a phrase or two after two
presentations. In the vast majority of my musical exams, I either just passed or failed on
these sections, so free recall is not a skill I have mastered. In music, free recall is an
interesting point of study when looking at the memory trace and the influences on and
products of the recall.
A study by Sloboda and Parker (2005) took eight young subjects, four who were
classified as musicians (trained performers) and four who were considered non-musicians
(they had not received any training but enjoyed listening to music). They listened to
excerpts from folk melodies. This study is interesting as they used real melodies as opposed
to pitch sequences designed specifically for the task, and is therefore more likely to show
how we react to and recall existing music. The excerpt was played six times and then the
subject was required to recall the sequence through singing six times. The authors found that
recalling more than one melody in a single session caused interference between the
materials, and so only the first melody recall was used in analysis. There was an intriguingly
high level of ‘intersong contamination’ (Sloboda and Parker 2005, 76), which was
unexpected.
The sequence was the first three phrases of the song ‘Sailor’, a Russian folk song
which was unfamiliar to the subjects. There were 21 notes, arranged in two-bar phrases, in
the form A1A2B. Section B shares much of the rhythmic patterns with A and so to an ear
attuned to Western tonality and structure, the melody would have been easily understood
and the structural groupings were clear. Transcriptions of the recalls were generated. Where
some subjects transposed their recall, it was ignored as unintentional and showed that
coding was in the context of pitch relationships, not actual pitch (Sloboda and Parker 2005,
77).
19
Firstly melodic contour was analysed. A melodic contour template was drawn up for
the three phrases and the subjects’ recalls were compared. Each correct note within a pattern
received one mark. When presented as a percentage, they found wide-ranging scores from
0-88%. Interestingly an example from one subject shows a very low melodic accuracy
score, but the subject seems to have created a new melody on the harmonic structure of the
original – this is analogous to linguistic recall, where it is possible to recall a sentence that
means the exact same thing but may contain very few of the same words: this new melody
essentially means the same thing (Bower 1976, cited in Sloboda and Parker 2005, 74).
Another subject’s recall seems to be based on the third phrase, which is section B, the
section containing different material to the previous two phrases. The third phrase recalled is
sometimes based on phrase one but the phrases have been rearranged. If the individual
phrase recalls were marked on the phrase they most resemble in the original, the score
becomes 57% as opposed to the original 2% (Sloboda and Parker 2005, 80). The subject's
initial error of recalling the third phrase first is maintained throughout the recalls, showing
how persistent a recall schema can be. This is most evident in the recalls of the other
melodies tested (which were not used in analysis), as the recalls were so disturbed by the
first melody that the materials became mixed through of the stubbornness of the recall
schema.
Looking at how a subject may improve over trials, it was found that while the notes
matching the melodies rose, the number of notes produced rose, too, meaning accuracy
stayed around the 50% mark. There is a constant freedom on the recall, which shows that
our recall is not highly restricted by the original surface material that we encode, suggesting
that it is not an absolute trace that we create. As discussed previously, our memory is much
more abstract (Peretz and Zatorre 2005, 97), based on underlying structures.
20
Is the A1A2B harmonic and melodic structure helpful in recalling the melodic
patterns? The similarity between the phrases must be measured. If the structure is preserved
in recall, then phrase three will be significantly less similar than phrase two to phrase one. In
the authors’ analysis of the original, it was indeed shown that phrase three was less similar
but in the results, it is seen that the subjects make phrase three more similar to phrase one
than in the original, but phrase two is still more similar to phrase one than phrase three. This
evidence points to the structure being reproduced, but that the recall is far less varied
structurally, which shows a tendency of memory towards simplification (Sloboda and Parker
2005 , 85).
A search between the two groups should show a difference between the musicians
and non-musicians. Musicians outperformed on all analyses except the consistency in
breathing places, yet interestingly the superiority was not always significant. In summary,
the recall of a melody was never note-perfect. Recalls were closely related to the original
melody in many ways but the best-retained element was the metrical structure. They all
retain a feeling of four in a bar, and this excellent performance on metre memory is
indicative of the primary use of metre as a structure for melodic comprehension and recall.
The phrase divisions (which is governed by harmony and metre) are also particularly well
adhered to, and the A1A2B structure is maintained even when the melody itself is incorrect.
However, the harmonic and melodic patterns are of varying degrees of accuracy throughout
the subject sample. It suggests that harmonic and melodic structures can be attained
independently of the rhythm, supporting Griffiths et al’s (1999, cited in Peretz and Zatorre,
2005; Fasanaro et al 1990) view that pitch and rhythm and pitch and metre may be separable
in brain function. And furthering this idea, it seems that harmonic structure can be separable
from melodic information (Peretz 1993). Several subjects’ recalls showed a very different
melody that fitted the harmonic structure, and it was common for rhythm, metre and
21
harmony to all remain intact while there were large variations on the melody. This retention
of harmony was the only significant difference between musicians and non-musicians,
suggesting training in Western music increases one's understanding of and memory for
harmonic and tonal structures (Sloboda and Parker 2005, 89).
This evidence suggests that we form models of structure in our memory where
surface detail is not retained (Peretz and Zatorre 2005; Sloboda and Parker 1985). Recall
involves filling in structural sections adhering to the 'rules' in that section. Different amounts
of training allow us to be sensible to the levels of structure. Due to this, the musicians can
code those harmonic relationships that are too complex for the non-musicians to
comprehend. However, there are certain elements that can be processed by both groups,
such as the basic melody and rhythm and the ascents and descents that are responsible for
the melodic contour. The main points of perception and memory for music that have been
discussed previously are here shown in action.
Chapter Summary In this chapter we have seen that the brain areas most implicated in the processing of
pitch are the right temporal cortex, in the anterolateral section, and particularly the superior
temporal gyrus as this contains the auditory cortex (Peretz and Zatorre 2005). Also
implicated are the left medial area and the splenium, an area deep inside the brain
‘underneath’ the auditory cortex. The junction between temporal, occipital and parietal
regions is also important, as damage here is correlated with a deficit in pitch processing
(Fasanaro et al 1990). In the right secondary auditory cortex, the anterior portion deals with
pitch chroma whilst the posterior portion deals with pitch height (Peretz and Zatorre 2005).
Melody is recalled as a series of pitch relationships, and the underlying structures such as
rhythm metre and harmony are much better recalled than the actual melody, even though the
melody is the most apparent factor in the music. Recall schemas are based on structures, not
22
the surface features and musical training increases your understanding and therefore
memory for these structures (Sloboda and Parker 2005).
Deutsch’s 1970 study shows that non-musical information doesn't interfere with
pitch, while musical analysis does, supporting the specialised system for the memory of
pitch and melody. Zatorre et al (1994) show us that working memory load activates auditory
areas in the superior temporal gyrus, meaning the auditory cortices are implicated in high-
load musical working memory, supporting the separate subsystem hypothesis also presented
by Gaab et al (2003). If there is indeed a musical memory subsystem, it is active in the
auditory cortex and the right temporal lobe, with longer-term musical memory leaning
toward action in the left temporal lobe. This side deals with language (Scott et al 2000) and
semantics (Koelsch and Siebel (2005) implicated the left anterior temporal lobe in their
study), and therefore musical semantics relating to our associations.
For AP possessors, there is reduced brain activity for working memory compared to
those not possessing AP. This is related to the categorical representation of pitch in AP as
opposed to a continuous trace for RP, which requires more maintenance of information.
When a single pitch is presented, AP possessors show action in the left posterior dorso-
lateral frontal cortex, which works with stimulus and response processes, in this case pitch
and pitch label. There is brain asymmetry in possessors suggesting that the smaller right-
sided auditory cortex may influence AP development.
23
Chapter 2 – Melody and Harmony
Melodic Contour Another method in which our brain processes melody is through the perception and memory
of melodic contour. The melodic contour is such a strong element in memory that a
reproduction of a non-Western melody, by a Westerner, would be produced with Western
elements, and yet the contour of the melody would be intact (Dowling 1978). Dowling
(1978) also points out that if the rhythm is incorrect or the intervals are displaced by an
octave (octave displacement: Deutsch (1972)) the tune may still be recognised. Deutsch
(1972) discovered when playing 'Yankee Doodle' with the notes randomly chosen from any
of three octaves, recognition was inhibited but still possible. She concluded that interval size
and preservation of contour are the most distinctive features for memory in our memory
trace. Long (1977) conjectured that a melodic contour with more directional change was less
effective in aiding the memory of pitch because more attention is required to the interval
size, where as a smoother, perhaps ‘V-shaped’, contour would create a shape memory trace
rather than intervallic memory trace.
The right superior temporal gyrus is strongly implicated in contour perception and
memory. Interval relations that make up the contour are processed in both the right and left
temporal gyri as discussed by Vignolo (2003), who found that the difference in accuracy in
the interval testing between those with right hemispheric lesions and those with left was
insignificant. Patel and Balaban in 1998 showed the cooperation between left and right
hemispheres in the processing of simple melody sequences (Peretz and Zatorre 2005, 92).
This asymmetry in the functions is already apparent in the tests on infants conducted by
Balaban et al (1998). The study found that reaction to a change that alters the contour of a
melody was stronger when presented in the left ear, while an interval change that preserved
the melodic contour provoked a stronger reaction when presented to the right ear (Balaban
24
et al 1998, 48). This is consistent with asymmetry in contour and intervallic change
discovered in adults but Anderson in 1994 (Balaban 1998, 48). This evidence suggests that
perception of contour may be stronger on the right-hand side of the brain (as the information
came from the left ear) and that memory for contour is stronger on the left-hand side. Once
again, encoding of both contour and intervals does not require the attention of the listener
(Tervaniemi 2003).
A study by Epinosa and Gernstein in 1988 used three-note series in all of their
permutations (ABC; ACB; BAC etc.) and found local networks of neurons processing these
series, and they discovered up to 24 neurons spiking simultaneously. The effect of one
neuron’s activity on another was specific to each sequence: neuron-spike patterns of one
tone changed depending on the preceding tone (ABC compared to BAC). Connections
between the tones depended on the pattern of the tones, which points towards contour being
coded at a neuronal network level (Weinberger 1999).
Scale Structure and Tonality The scale structure involved in a tonal melody is a useful tool for listeners in the perception
and memory in pitch sequences and it is often learnt implicitly and used automatically to
influence incoming pitch, melody and interval information (Tillmann and Bharucha 2000).
Tonality can provide a ‘cognitive landmark’ (Snyder, 2009) and when compared with non-
musicians, although musicians' recall of music is superior, Cook found in 1987 that it is less
superior when they are presented with an atonal sequence (Lehmann et al 2007). Brain
damage can inhibit tonal knowledge while the interval and contour perception are spared
(Peretz 1993). In a case study by Peretz (1993) the subject who had a lesion in the left
temporal lobe and right frontal-opercular region (which includes the right temporal gyri)
could no longer judge the difference between tonality and atonality, and had a reduction in
pitch memory, but could still perceive the pitches and melody contour. There may be
25
specialised neurons for processing of scale structure because it can be separable from other
functions. Janata, Birk, Van Horn, Leman, Tillmann and Bharucha (2002) suggested that the
medial prefrontal cortex was associated with knowledge of the key and tonality of a piece,
maintaining a tonality map. The authors discovered that the connections to the medial
prefrontal cortex were mainly coming from areas associated with auditory processing, and
that this auditory task-related activity was seen mainly in the right hemisphere of the
anterior superior temporal gyrus (Janata et al 2002).
Long (1977) conducted an experiment whereby he tested the effects of melody
length, contour, tonal structure and musical ability on memory for pitches within melodies,
which is different to other studies testing a sequence recall ability. Unsurprisingly, two of
the three groups that were comprised of mostly music students performed far better than the
one group consisting of mostly non-music students, supporting the view that musical
training would increase memory for pitch. Memory may be dependent on the learned
musical systems. He found that the length was not significant but that the tonality of
melodies, as opposed to atonality of others, increased the ease at which the pitch was
remembered. The learned system of tonality was a great aid to this memory (Long 1977).
A knowledge of tonal and harmonic principles lead to expectancies in harmonies.
Tramo, Bharucha and Musiek studied brain damaged subjects in 1990 and found that they
retained the ability to generate chord expectancies after a bilateral lesion in auditory cortex
but were deficient in pitch processing and had difficulty in judging intonation (Peretz and
Zatorre 2005, 93). Deviation from the chord expectancies causes robust potentials in the
brain, in the inferior frontal areas where the neurons generate a response. Koelsch, Gunter,
van Cramon, Zys-set, Lohmann and Friederici in 2002 showed that this happens bilaterally
(Peretz and Zatorre 2005, 93). The right hand side is typically implicated in yes/no decisions
and perhaps in inhibition responses to risk, which may be linked to the surprise of a
26
harmonic deviation. The left hand-side is used in language functions and grammar (Scott et
al 2000), which corresponds to the ‘grammar’ of music, and the change in direction for the
music. Our grammar of music however, is limited to the musical culture in which we are
saturated. A Western-cultured person would reproduce a melody of entirely non-Western
elements inside the Western tonality schemes (Dowling 1978). Bartlett and Dowling in 1980
(Shuter-Dyson 1999) saw a recognition of key change, when the key was distant, in children
as young as 5 whilst children aged 9-10 showed a 90% accuracy in identifying a tonality
change.
Consonance and dissonance naturally play a part in our perception and recognition
of pitch relations. Intervals that are consonant have a simple frequency ratio (for example,
the fifth is 3:2) and dissonant relations have more complex frequency ratios (16:15 for the
minor second). The superior temporal gyri are again implicated here as dissonance appears
to be computed bilaterally in this area, possibly by specialised mechanisms. Fishman,
Volkov, Noh, Garell and Bakken in 2001 found dissonance caused activity in the temporal
gyri, but not consonance (Peretz and Zatorre 2005, 94), and a bilateral lesion of the auditory
cortex (inside this gyrus) has the effect of a loss in the subject's sense of dissonance (Peretz
et al 2001, in Peretz and Zatorre 2005, 94).
Some questions brought up here by Peretz and Zatorre (2005) are that of the point in
perception where computation of dissonance is critical, and whether our sensitivity is to
dissonant noises in general or to dissonant pitch relations. Is our definition of dissonance
down to our natural auditory-musical functions, or is it through learned associations? Lynch
and Eilers in 1991 tested western children's judgment of dissonance in western and Javanese
music, and found that while the western music was more accurate, they performed better
than the chance accuracy (predicted accuracy if purely guesswork) (Carterette and Kendall
27
1999, 731), so there is some argument for a innate sense of dissonance, but with greater
sensitivity in the familiar and learned contexts.
Musical Notation and Structure Tonal structure's role in memory has already been discussed, which has its largest effect in
aural presentation. Does the visual representation have an effect on our memory similar to
that of the musical structure? Standard notation represents musical pitch and rhythm in a
visual manner. Ordered and 'good' musical structure is remembered far more easily than
'bad' or random structure (Halpern and Bower 1982).
Halpern and Bower used an interesting experiment on chess experts and chess
amateurs by Adriaan de Groot in 1965 showing results in how they perceive and remember
patterns. He found that to remember patterns and moves both during and after play, chess
Grand Masters use chunking, a memory method involving grouping information together to
create one 'chunk', rather than several smaller ones. The amateurs did not do this. It suggests
that other experts, such as musicians highly skilled in reading of notation, would also use
chunking in perception (in tasks like sight-reading) and memorisation while amateurs would
not. However, de Groot found that the experts’ memory is superior in meaningful patterns,
but no better than the amateurs’ when the patterns were random. Halpern and Bower (1982)
believed it would be the same for musicians when musical patterns and structures were
ordered and meaningful through a visual presentation.
In Halpern and Bower’s experiment they used sets of ten crotchets with no
accidentals or repeated notes and they were placed into pairs of 'good' structures and 'bad'
structures with very similar contours and intervals, meaning they are visually similar. These
melodies were judged to be 'good' or 'bad', both visually and aurally, by separate groups of
musical graduates and expert performers. The 'good' were judged significantly better than
the 'bad', which confirms their classification on both visual and aural presentation. They
28
were also judged equally in visual complexity by non-musicians. 'Random' melodies were
composed in addition to the 'good' and 'bad' with larger intervals and more contour changes.
In the pilot study they discovered some possible encoding methods for the notation. Halpern
and Bower had a group of twelve musicians and twelve non-musicians. A slide of a 'good' or
a 'bad' melody was displayed for five seconds, and then recalled after a delay of fifteen
seconds. They found that musicians were overall better at recall than non-musicians and that
the 'good' melodies were overall remembered far more accurately than the' bad'. Whilst the
musicians remembered the 'good' melodies more than the 'bad', the non-musicians
performed equally poorly on both the categories, which supports the judgement that the
melodies were equally complex visually. The musicians’ memory for the 'bad' melodies was
better than the non-musicians’ performance on the 'bad'
In the second experiment, Halpern and Bower (1982) further investigated the effect
of the musical structure on the notation. Since the musicians recalled 'bad' melodies better
than the non-musicians, what kind would they recall equally? Random note placement in the
random melodies causes a far greater interval size average, which leads to an increase in
complexity of the visual presentation. This should affect both groups as while the non-
musicians only use visual regularity, musicians use both visual and musical regularity. The
random note placement inhibits the musical regularity as well as the visual. The musicians
showed that they performed well for 'good' melodies as they had both good visual and
musical regularity, moderately for the 'bad' melodies, which also has good visual regularity
but poor musical regularity, and badly on the random melodies that lacked all regularity.
The non-musicians performed moderately, and equally so, on the 'good' and 'bad' melodies
and badly on the random melodies. The musical differences had no effect on the non-
musicians, only the visual differences, and so they were worst performers on the random
melodies as there was little visual structure. Musicians were only mildly superior to the non-
29
musicians on the random melodies, as there was no musical or visual regularity to follow,
which means melody had to be remembered in terms of pitch names. This was unfamiliar to
the non-musicians, probably explaining their poorer performance.
Absolute accuracy on recall was initially scored, but this did not allow for near
misses (like placing a note one line too far down on the stave) or for meaningful musical
mistakes (such as octave displacement), so the memory for the overall visual representation
of notation was difficult to judge. Instead it was marked on similarity – a melody placed
consistently one note too low on the stave in recognition would have been marked zero on
absolute, but actually it is very close to perfect so the memory was very good. This was
marked out of ten. The non-musicians scored 2-3/10 and the musicians’ score was far more
wide-ranging, but none were approaching perfect. The musicians’ scores were as predicted,
but their advantage over non-musicians decreased from 'good' to 'bad' to random. Their
similarity points were far higher than the non-musicians'. Musicians made some musical
errors. Often when memory was not intact, musically sensible substitutes were inserted. In
other examples, notes were encoded as an F major triad, for example, but recalled in the
wrong order. This is an example of the chunking of patterns within melodies. There were
fewer of these mistakes in the less structured 'bad' melodies as compared to the 'good', as
there was less musically structural information to encode.
Even when the recall was immediate, the musicians were better at recalling the
'good' melodies, which shows the immediate use of the learned structures as a framework
for memory. The effect of musical structure on working memory for pitch is shown in sight-
reading experiments by Sloboda (cited in Gabrielsson 1999, 512) where musicians
processed the music as a series of groups, both structural and musical. After a very brief
exposure to melody, the musicians were superior in recall of the simple patterns, and
reported that they seemed to be obvious in their patterns and groups.
30
Two characteristics are likely to be helpful in memory for pitch sequences. Firstly,
the ease with which you can group the melody into sub-patterns and secondly the number of
patterns that are likely to be formed. Following the theory that musicians will use chunking
to remember patterns, the more easily sectioned melodies should be easier to remember, and
even easier if there are fewer chunks of information to maintain. It is expected that chunk
size is related to expertise (Lehmann et al 1007, 112). In a well-structured melody,
groupings would be easier and fewer, and thus more easily remembered than poorly
structured 'bad' melodies. Visual cues are unlikely to indicate the groupings, so non-
musicians, who are relying on the visual structures, are unlikely to use this method of
memory for recall.
Using the same 'good' and 'bad' melodies from the previous tests, Halpern and Bower
(1982) presented these melodies on a sheet of paper in a random order to a group of subjects
(seven musicians and seven non-musicians) and instructed them to use vertical lines to
group the melodies as they wished. Consistency of absence or presence of a particular
division for both groups was marked along with the average number of groups for each
melody and the number of divisions each subject gave each melody (Table 3). Interestingly,
the difference between the two groups on consistency in the 'good' melodies was
insignificant, and in fact it was the non-musicians who were more consistent in their choices
for groupings. However, the non-musicians' choices seemed not to be musical decisions and
the consistency came from decisions like a visual division where the direction of the note-
stem changed. Another interesting finding was that the 'good' melodies were no more
consistently grouped than the 'bad' melodies, for either group, which shows that the musical
structure which makes it 'good' is not a visually obvious element, and therefore is not seen in
the 'good' melodies by the non-musicians, and that a musical division may be subjective and
differently judged by different musicians.
31
To see whether the musicians were using the chunking strategy, recall scores from
immediate recall of melodies were compared with the grouping consistency marks for each
melody. As expected, the non-musicians showed no correlation between their grouping
consistency and accuracy in immediate recall whilst there was a significant correlation for
the musicians, pointing to a suggestion that chunks that are clear in the melody are used to
ease the memory process when a quick glance is the only presentation.
When there are fewer chunks, there is less information to maintain. Musicians
marked a significantly lower number of musical groups or chunks in the 'good' melodies
than the 'bad' melodies, but for the non-musicians it was the other way around. This was
explained after exploring the number of stem direction changes for each group of melodies
and the 'bad' melodies had fewer changes than the 'good' melodies. This shows the
importance of the musical structures for the musicians over the easier visual structures like
the grouping of the stem changes. The non-musicians were relying heavily on the visual
grouping strategies over musical groupings, as they showed a very high correlation between
the individual grouping numbers for each melody and the mean average throughout the
whole group. This suggests they used the same stimulus to mark segmentation in the
melody. The average number of groupings for each melody was then compared with the
recall marks and there was no correlation for the non-musicians. Clearly the stem direction
segmentation was not a method that was implemented by or that helped the non-musicians'
Table 3: Mean
consistency score
and number of
groupings for good
and bad melodies.
Halpern and Bower
(1982).
32
recall performance. The musicians showed a significant negative correlation between
number of groupings and performance in both delayed and immediate recall (in which it was
even more significant), meaning they made fewer groups and therefore performed better.
The musicians grouped musically, not visually, and could take advantage of these groupings
in memory.
The inconsistency in the musicians’ groupings may have come from the lack of
musical features other than pitch. Segmentation, patterns and phrases are usually indicated
by larger scale contours and rhythms, neither of which are present in these melodies. The
fact that they ignored the visual cues of the stem directions (whilst the non-musicians took
advantage of this and consequently were more consistent) can explain the lack of
consistency through the group.
These experiments showed the negative correlation between the number of chunks
and recall accuracy in musicians' recall and therefore that chunking was a tool used by the
musicians, knowingly or otherwise, in order to recall melody. These relations are stronger
when the recall is immediate which, along with the findings that well-structured melodies
aid recall, show the very important and fast influence on memory trace that musical
structure can have, suggesting very early use of 'music-specific encoding' (Halpern and
Bower 1982, 46). It also shows that the non-musicians do not use these strategies, which
suggests either an innate brain function or an element of training is involved.
The 'grammar' of music (Halpern and Bower 1982, 47) has been learnt by the non-
musicians (to some extent) through listening to music and this is clear from the inter-
participant reliability of the non-musicians judging whether the melodies were 'good' or
'bad'. They were relatively consistent in deciding which were 'good' and 'bad', even though
the 'bad' were not 'blatantly discordant' (Halpern and Bower 1982, 47). Musicians have also
33
gained this understanding and built upon it through their training, strengthening their
expectations and knowledge of structures and activities. This includes the names of relations
and structures, which can aid their memory for these features and elements of music. The
knowledge of the grammar involved in a 'good' melody and their associative knowledge of
how it is presented visually means recognition of groupings, patterns and structures occurs
immediately. Patterns which are well known (such as triads, or rising scales) can aid the
expectation of what is to come, and this would explain the existence of musically sound
mistakes in recall of melodies - the errors are not haphazard and come from interference
from similar musical patterns stored in the long term memory (Sloboda 1978). A melody is
presented as a ‘hierarchy of subsequences’ (Deutsch 1980, 2) and structure is utilised in
recall. Deutsch also found that the unstructured sequences caused a much heavier memory
load. The expectancies are quickly confirmed and the tonal hierarchy also helps us to
organise these chunks to aid the memory (Deutsch 1980). The chunks and musical structure
are often interlinked so when one strategy fails in retrieval, another can be used (Lehmann et
al 2007, 120).
Some studies suggest that the music is not stored in a simply visual or aural form but
in a more abstract manner recording contour (Dowling 1978) and tonality (as this is a well-
defined mental structure on which to place the melody) (Dowling 1978) along with other
musical elements. Sloboda’s 1978 example of 'chunks' is the sequence [C-D-E-F-G-D-B-C],
coded as:
a) C scale rising to dominant;
b) falling dominant chord;
c) return to tonic.
A novice musician unaware of tonal structure might encode:
a) C up to G;
b) G down to D;
34
c) D down to B;
d) B up to C.
This comprises of more chunks and they have no connecting link between then. A
non-musician might then code the contour and approximate notation, as he has no access to
abstract frameworks. One builds an expectation on what you can see, and with confidence in
your training you may not need to consciously look at all the information (Lehmann et al
2007, 116; Tervaniemi 2003). Halpern and Bower (1982) suppose that the musicians use
these abstract methods and musical names rather than attempting to 'hear' the music in their
heads as this might cause interference (Sloboda and Parker 2005) as one would be
essentially sight-reading the presented melody, not concentrating on the memory trace.
Chapter Summary This chapter has shown the importance of preservation of contour for memory and for
recognition. The right superior temporal gyrus is activated in both perception and memory
for melodies and contour. The right brain seems to accommodate contour change while the
left preserves and recognises contour. The connections (interval change and direction)
between the pitches cause neurons to behave differently, which brings contour to a neuronal
level of perception, meaning it is a very detailed part of the function.
Tonal structure seems to be learnt automatically and acts as an aid in memory for
music. From brain damage studies we saw that a bilateral lesion in the temporal lobe led to a
loss of pitch and tonality recognition, but the tonal processing function remained, suggesting
that musical perception and memory belong to separable systems. A bilateral lesion also led
to a deficit in pitch processing but the ability to generate chord expectancies was retained,
showing that harmonic systems may be separable from the melodic systems. The right
anterior superior temporal gyrus is connected to knowledge of tonality (as well as working
memory) and so must be instrumental in sight-reading. The inferior frontal region is
activated in harmonic deviance bilaterally – on the right side because of surprise (linked to
35
risks) and the left for the analysis of change in 'grammar'. Bilaterally, the superior temporal
gyrus is activated in dissonance – the right side for musical sound and the left for grammar
or semantics.
The tonal structure plays a big part in a musician's memory for music, and while the
visual structure can help (as 'bad' melodies were recalled better than 'random' melodies in
Halpern and Bower's 1982 experiment) it is secondary to music. The grouping of the
melodies showed that chunking, expectations and tonal knowledge are all methods
employed to remember a melodic sequence.
36
Chapter 3 – Music in Action
Sight-Reading and Working Memory Sight-reading is an important musical activity, which involves very brief and complicated
processes in the brain. The sight-reading of music is functionally and neurally different to
our reading of words. For example, in some brain damage studies Cappelletti, Waley-
Cohen, Butterworth and Kopelman in 2000 found that symbol reading can remain intact
while musical notation reading is inhibited (Peretz and Zatorre 2005, 101). fMRI scans
revealed the involvement of the right occipital temporal region in deciphering musical
notation (Schon et al 2000, in Peretz and Zatorre 2005, 101). Sight-reading requires vast
amounts of information to be processed in a very short time for temporary use. One must
interpret both pitch and rhythm, sometimes on more than one stave, comprehend and
implement a key signature and metre and recognise the familiar patterns. Often these are
perceived peripherally, meaning errors that occur are 'musically meaningful' as they fit into
the perceived structure (Sloboda 1977, in Sloboda 2005, 40). The optimum time span for the
working memory to process music is 5-8 seconds and the surface detail is therefore partly
set-aside in order for the larger-scale structure to be comprehended (Snyder 2009). A
performance plan for motor translation must be created and the brain must anticipate correct
continuance of the piece in order to recognise mistakes automatically, using expectations
and imagery.
As has been discussed, some of these actions are distinguishable neurophysically.
After brain damage, pitch and rhythm perception can become separated (Fasanaro 1990).
There can be damage at different levels of perception and decoding, even as early as the
visual perception where pitch and rhythm are presented in different manners. It may happen
later on, when the brain attempts to create an imagined representation. Pitch is encoded
differently for different people. Some may give note names while others would use a motor
37
function (like pressing the key on a piano) so it could be one of these functions that is
damaged (Peretz and Zatorre 2005, 102). However, since many can sight-read in different
media, through singing or several instruments, representation of pitch is likely to be quite
abstract and unrelated to the medium of performance. Koelsch and Siebel (2005) implicated
the frontal, parietal and premotor cortical areas in working memory, along with the
cerebellum.
According to McPherson in 1993, sight-reading is one the five skills all musicians
should possess (Kopiez et al 2006). Activities and skills relevant to sight-reading are aural
imagery, sight reading experience, your style of thinking and possession of an external locus
of control (Kornicke 1995, in Kopiez et al 2006). Sloboda (1974) noted that one’s ability to
read ahead is a condition and a sufficient short-term memory buffer must be present in order
to read ahead whilst playing the previous music. He also suggested that being left-handed
may be an advantage as left-handed people generally have less of a handicap on their right-
hand side than right-handed people do on their left side. Kilshaw and Annett in 1983 also
found that musical imagery appears to be highly related to sight-reading, as efficiency in
this department may help to hear the music and understand how it should sound before it is
executed, all within the small space of time (Kopiez et al 2006).
Imagery and Mental Rehearsal Musical working memory is essential for musical imagery, described as imagining music or
its attributes without hearing it (Peretz and Zatorre 2005, 97). Perceptual defects and
imagery defects are often found in common when a lesion in the auditory cortex is present,
which supports the view that imagery uses the perceptual mechanisms. In some tests with
brain imaging, it was discovered that the secondary auditory cortex is activated in both
imagery and rehearsal of melodies (Halpern and Zatorre 1999). They confirmed their
predictions that the right associative auditory cortex (which provides a realistic experience
38
for the brain) accompanies musical imagery, and that the imagery cues retrieval from the
‘musical semantic memory’ (Halpern and Zatorre 1999, 702) and this increases CBF in the
right inferior frontal region and bilaterally the middle frontal region. When there is no need
for ‘semantic retrieval’ then the left frontal area is recruited. However, it is the associative
auditory cortex, not the primary, which is active in the imagery. They used melodies both
with and without lyrics. Those with lyrics activated the temporal cortex bilaterally whereas
those without activated the right temporal cortex. The right auditory cortex activation
supports the hypothesis that the right hemisphere mechanisms process the tonal information
(Halpern and Zatorre 1999, 702).
Highben and Palmer in 2004 found when they removed the auditory feedback of
rehearsal and requested the participants to imagine the feedback, that performance was
poorer than after rehearsal with the auditory feedback. This shows that the auditory
feedback is important for a successful performance. The pianists with excellent aural
imagery were least affected by the removal, supporting the view that aural imagery involves
the same brain sequences as actual aural information (Palmer 2005).
Janata (2001) found that the imagination of the continuance of a melody (where they
omitted an unexpected note) showed similar electrical functions in the brain to when
perception of a real tone occurs, which suggests that the auditory mechanisms which are
responsible for perception are also involved in imagery (Halpern and Zatorre 1999).
The neural networks involved in musical ability can be studied by teaching non-
musicians to play music (Peretz and Zatorre 2005). Musically untrained subjects were
allocated two hours of piano for five days, which, besides resulting in vast improvement in
performance, also resulted in increased activity in the motor areas associated with hand
muscles. Primary motor cortex neurons also showed increased activity, and repeated
39
information input to the motor cortex resulted in long-term potentiation of the motor
neurons (Pascual-Leone 2001). Interestingly mental practice (‘the imagined rehearsal of a
motor act with the specific intent of learning or improving it’ (Pascual-Leone 2001, 321))
has similar effects:
‘Studies of regional Cerebral Blood Flow (rCBF) suggest that the prefrontal and
supplementary motor areas, basal ganglia, and cerebellum are part of the network
involved in the mental simulation of motor acts. Therefore, mental simulation of
movements activates some of the same central neural structures required for the
performance of the actual movements. In so doing, mental practice alone seems
to be sufficient to promote the modulation of neural circuits involved in the early
stages of motor skill learning. This modulation not only results in marked
improvement in performance, but also seems to place the subjects at an
advantage for further skill learning with minimal physical practice. The
combination of mental and physical practice leads to greater performance
improvement than does physical practice alone, a phenomenon for which our
findings provide a physiological explanation" (Pascual-Leone 2001, 321).
This shows the quick effects of training on brain plasticity. There are several
differences therefore in the way that musicians and non-musicians use the musical
mechanisms in the brain and listen to music. It seems that musicians recruit more neural
tissue for the activities and are more analytical in their listening. However, this is not
exclusive to musicians as non-musicians have shown that they too consider interval size
both consciously and automatically (Tervaniemi 2003). Even so, Peretz and Babai in 1992
found that musicians are more flexible in their listening, using their learned tonal and
structural knowledge to decide which method would be most appropriate for coding,
analysis and memory-trace generation (Peretz and Zatorre 2005, 105).
40
Brain Function and Structure It is clear there is right-sided asymmetry for pitch-based processing but music uses a vast
network of regions on both right and left sides of the brain. Music and the activities
involved are both complex and numerous and it is not clear which part of the brain are
music-specific. Many of them share neurons and systems with other functions. For example,
Koelsch (2002) found that harmonic deviation activates Broca's area inside the frontal lobe,
which is a region associated with language, suggesting an element of syntactical procedure.
However, the area is vast, and could easily accommodate several networks, meaning the
musical function and the language function could be entirely unrelated in this area (Peretz
and Zatorre, 2005). The brain's structural adaptions to each function happen at a gross
structural level and appear during early-aged training (Altenmueller and Schneider, 2009).
Chapter Summary The studies in this section have shown that the auditory cortex is implicated in the imagery
of music and in working memory with the frontal, parietal and premotor cortical areas and
the cerebellum. Imagery needs the perceptual mechanisms in order to occur, like mental
rehearsal, which involves the primary motor cortex and the right associative auditory cortex
is involved in musical imagery. For musical semantic retrieval the right inferior frontal
region and bilateral mid-frontal region are used. When lyrics are used in perception, the left
and right auditory cortices are both involved as the left deals with language while right deals
with sound (in this case tonal information), and when without lyrics, only the right side is
used. In comparison with actual aural feedback, imagery causes the same activation in the
brain, and similarly when imagining a continuing melody, the same activation appears as
when actual perception is occurring.
Many separable processes are involved in sight-reading, but for them to work
simultaneously, one must have an ability to read ahead, an adequate short-term memory
41
buffer and knowledge of the musical tonal and structural and frameworks to aid imagery and
generate accurate expectations.
42
Conclusion
During this study, many activities and processes have been addressed in order to understand
the human brain’s means of dealing with musical information. It has led me to some
conclusions about my own patterns of perception, analysis and memory.
I may have been more able than my peers to acquire musical skills because my entire
family is musical. It is not necessarily genetic, but environmental. Studies have shown that
much musical information can be processed automatically, and therefore knowledge can be
unknowingly assimilated. Being surrounded by music in my very early years and onwards
must have had an effect on the development of musical structures and functions in my brain.
My ability to sight-read is likely due to a combination of things, but I believe the
main one may be an efficient short-term working memory buffer. This allows me to process,
perceive and reproduce the many facets of music in a very short space of time, and its
efficiency allows me to do this accurately. I have found in other areas of life, I am very able
to ‘juggle’ tasks: I can participate in several conversations, talk and read simultaneously and
even in more academic situations like mathematics, several activities were easily possible.
In comparison to my cognitive style, my sister cannot concentrate on more than one task,
but can concentrate for much longer. During aural recall, she can keep her mind on the
presented material without thinking about other things, and there are fewer facets to
perceive in this fashion of presentation. She is able to hold the information and recall it with
better accuracy. However, she cannot comprehend all the information involved in sight-
reading fast enough to reproduce the music accurately. A secondary reason could be an
aptitude in musical imagery. As I am very able to “hear” the music as I read it, it will aid my
understanding of the piece, and help to build those expectancies built on tonal frameworks.
When required to sing back a presented melody, I am aware that whilst the contour
of the recalled melody is usually correct (as was the case in Sloboda and Parker’s 2005
experiment), the tonal structure often is not. As I don’t have absolute pitch, my analysis of
43
pitch relations and intervals to generate a trace, but these are based inside a tonal structure,
and it appears that the frameworks that I have understood and maintained are not being used
to aid my memory, unlike perception during sight-reading.
I think my abilities come down to my preference for the visual presentation of
music, as opposed to aural presentation. Essentially, aural recall and sight-reading amount to
the same processes – perception and reproduction. However, aural recall requires fast
perception, memory and recall of aural information, whereas sight-reading requires fast
perception of a larger amount of information through written music followed by
reproduction. Learning a piece of music by ear is impossible for me, and I find it incredibly
difficult to perform the most basic analysis without reading the score. In other tasks, such as
learning languages, I could pronounce and remember the word if I saw it. In an informal test
on myself, I could remember musical phrases such as in Halpern and Bower’s experiment
(1982) through seeing the musical patterns and chunks – a combination of the musical and
the visual methods. I think perhaps my difference in skill is not due to musical memory or
analysis, but a failing in aural perception, and a balancing superiority in visual perception.
Many brain areas, methods, processes and contextual effects (such as mood at the
time of testing) have an effect on performance and skill in music, so it is impossible to
pinpoint the reason for the difficulties I and other musicians have in certain activities, but
through knowledge of my own processes and through this study of perception and memory,
I think I have uncovered a path towards the answers.
44
References Altenmueller, E. and S. Schneider (2009) ‘Planning and performance’ in Oxford Handbook
of Music Psychology. S. Hallam, I. Cross and M. Thaut (eds.) pp 332-343, Oxford:
Oxford University Press.
Baharloo, S., Susan K. Service, Neil Risch, Jane Gitschier and Nelson B. Freimer (2000)
Familial Aggregation of Absolute Pitch. The American Journal of Human Genetics, 67,
755–758.
Balaban, M. T., Linda M. Anderson and Amy B. Wisniewski (1998) ‘Lateral Asymmetries
in Infant Melody Perception’. Developmental Psychology, 34, 1:39-48.
Bermudez, P. and Robert J. Zatorre (2009) ‘The absolute pitch mind continues to reveal
itself’. Journal of Biology, 8:8, 75.
Burns, E.M. (1999) 'Intervals, Scales, and Tuning' in The Psychology of Music D. Deutsch
(ed.) pp. 251-264, London: Academic Press.
Carterette, Edward C. and Roger A. Kendall (1999) ‘Comparative Music Perception and
Cognition’ in Psychology of Music D. Deutsch (ed.). 725-791, Academic Press: London.
Deutsch, D. (1970) ‘Tones and Numbers: Specificity of Interference in Immediate
Memory.’ Science 168, 1604-1605.
Deutsch, D. (1972) ‘Octave generalization and tune recognition.’ Perception &
Psychophysics, 11, 411-412.
Deutsch, D. (1973) ‘Octave generalization of specific interference effects in memory for
tonal pitch.’ Perception and Psychophysics, 13:2, 271-275.
Deutsch, D. (1980) ‘The Processing of Structured and unstructured tonal sequences.’
Perception and Psychophysics, 28, 381-389.
Deutsch, D. (1999) ‘The Processing of Pitch Combinations’ in The Psychology of Music, D.
Deutsch (ed.) pp 349-412, Academic Press: London
Dowling, W. Jay (1978) ‘Scale and Contour: Two Components of a Theory of Memory for
Melodies.’ Psychological Review, 85:4, 341-354.
Fasanaro, A. M., D. L. A. Spitaleri, R. Valiani and D. Grossi (1990). ‘Dissociation in
Musical Reading: A Musician Affected by Alexia without Agraphia.’ Music Perception:
An Interdisciplinary Journal, 7:3, 259-272.
Gaab, N., Gaser, C., Zaehle, T., Jancke, L., Schlaug, G. (2003) ‘Functional anatomy of pitch
memory - an fMRI study with sparse temporal sampling.’ NeuroImage, 19, 1417–1426.
Gabrielsson, A (1999) ‘The Performance of Music’, in The Psychology of Music, D.
Deutsch (ed.), pp 501-602, Academic Press: London.
45
Halpern, A.R. (1989) ‘Memory for the absolute pitch of familiar songs.’ Memory &
Cognition, 17:5, 572-581.
Halpern, A. and G. Bower (1982) ‘Musical Expertise and Melodic Structure in Memory for
Musical Notation.’ The American Journal of Psychology, 95:1, 31-50.
Halpern, A. and Robert J. Zatorre (1999) ‘When that tune runs through your head: a PET
investigation of auditory imagery for familiar melodies.’ Cerebral Cortex, 9, 697-704.
Houston, D. and Geoffrey Haddock (2007) ‘On auditing auditory information: the influence
of mood on memory for music.’ Psychology of Music, 35:2, 201-212.
Janata, P. (2001) ‘Brain Electrical Activity Evoked by Mental Formation of Auditory
Expectations and Images.’ Brain Topography, 13:3, 169-193.
Janata, P., Jeffrey L. Birk, John D. Van Horn, Marc Leman, Barbara Tillmann, Jamshed J.
Bharucha (2002) ‘The Cortical Topography of Tonal Structures Underlying Western
Music.’ Science, 298, 2167-2170.
Keenan, J.P., Ven Thangaraj, Andrea R. Halpern and Gottfried Schlaug (2001) ‘Absolute
Pitch and Planum Temporale.’ NeuroImage, 14, 1402–1408.
Koelsch, S. and Walter A. Siebel (2005) ‘Towards a neural basis of music perception.’
TRENDS in Cognitive Sciences, 9:12, 578-584.
Kopiez, R., C. Weihs, U. Ligges, and J.I. Lee (2006) ‘Classification of high and low
achievers in a music sight-reading task.’ Psychology of Music, 34:1, 5-26.
Lamont, A. and Nicola Dibben (2001) ‘Motivic Structure and the Perception of Similarity.’
Music Perception, 18:3, 245-274.
Lehmann, A., J. Sloboda and R. Woody (2007) Psychology for Musicians. Oxford: Oxford
University Press.
Long, A.P. (1977) ‘Relationships between Pitch Memory in Short Melodies and Selected
Factors.’ Journal of Research in Music Education, 25:4, 272-282.
Miyazaki, K. (1992) ‘Perception of Musical Intervals by Absolute Pitch Possessors.’ Music
Perception: an Interdisciplinary Journal, 9:4, 413-426.
Powerhousemuseum.com (2001) Cochlear implant. [online] Available at:
http://www.powerhousemuseum.com/hsc/cochlear/hearing.htm [Accessed: 27 Apr 2013]
Palmer, C. (2005) ‘Sequence Memory in Music Performance.’ Current Directions in
Psychological Science, 14:5, 247-250.
Pascual-Leone, A. (2001) ‘The brain that plays music and is changed by it.’ Available at:
http://tmslab.org/wp-content/files/PascualLeone_MUSICBRAIN_NYAcadSci.pdf
[Accessed: 26 Apr 2013].
Peretz, I. (1993) ‘Auditory Atonalia for Melodies.’ Cognitive Neuropsychology, 10:1, 21-56.
46
Peretz, I. (1996) ‘Can We Lose Memory for Music? A Case of Music Agnosia in a
Nonmusician.’ Journal of Cognitive Neuroscience, 8:6, 481-496.
Peretz, I. & Robert J. Zatorre (2005) ‘Brain organization for music processing.’ Annual
Review of Psychology, 56, 89-114.
Pierce, J. R. (1999) ‘The Nature of Musical Sound’ in The Psychology of Music Deutsch, D.
(ed.) pp. 1-24 London: Academic Press
Sacchetti, E., Alessandro Galluzzo, Adelaide Panariello, Giovanni Parrinello and Stefano
Francesco Cappa (2008). ‘Self-ordered pointing and visual conditional associative
learning tasks in drug-free schizophrenia spectrum disorder patients.’ BMC Psychiatry,
8:6.
Scott, S.K., C.C. Blank, S. Rosen and R.J.S. Wise (2000) ‘Identification of a pathway for
intelligible speech in the left temporal lobe’. Brain, 123, 2400-2406.
Shuter-Dyson, R. (1999) 'Musical Ability' in Psychology for Musicians D. Deutsch (ed.) pp
627-651, London: Academic Press.
Skidmore.edu (n.d.) Auditory Stimulus & Anatomy. [online] Available at:
http://www.skidmore.edu/~hfoley/PercLabs/Auditory1.htm [Accessed: 26 Apr 2013].
Sloboda, J. (1974) ‘The Eye-Hand Span – An Approach to the Study of Sight-Reading.’
Psychology of Music, 2:2, 4-10.
Sloboda, J (1978) 'The Psychology of Music' in Exploring the Musical Mind J. Sloboda
(ed.), 3-26, New York: Oxford University Press.
Sloboda, J (2005) Exploring the Musical Mind. New York: Oxford University Press.
Sloboda, J and D.H.H. Parker (2005) 'Immediate Recall of Melodies' in Exploring the
Musical Mind J. Sloboda (ed.), 71-96, New York: Oxford University Press.
Snyder, B (2009) 'Memory for Music' in Oxford Handbook of Music Psychology. S. Hallam,
I. Cross and M. Thaut (eds.), pp. 107-118, Oxford: Oxford University Press.
Stewart, L., K. von Kriegstein, J.D. Warren, and T.D. Griffiths (2006) ‘Music and the brain:
disorders of musical listening.’ Brain, 129, 2533-2553.
Tervaniemi, M. (2003) ‘Musical Sound Processing: EEG and MEG evidence’ in The
Cognitive Neuroscience of Music Peretz and Zatorre (ed.), pp. 294-309, Oxford: Oxford
University Press.
Tillmann, B. and Jamshed J. Bharucha (2000) ‘Implicit Learning of Tonality: A self-
organizing approach.’ Psychological Review, 107:4, 885-913.
Vignolo, Luigi A. (2003) ‘Music Agnosia and Auditory Agnosia: Dissociations in Stroke
Patients.’ Annals of the New York Academy of Sciences, 999, 50-57.
47
Weinberger, Norman M. (1999) ‘Music and the Auditory System’ in The Psychology of
Music Deutsch, D. (ed.), pp. 47-89, London: Academic Press.
Zatorre, R.J., Alan C Evans and Ernst Meyer (1994) ‘Neural Mechanisms Underlying
Melodic Perception and Memory for Pitch.’ The Journal of Neuroscience, 14:4, 1908-
1919.
Zatorre, R. J., David W. Perry, Christine A. Beckett, Christopher F. Westbury and Alan C.
Evans (1998) ‘Functional anatomy of musical processing in listeners with absolute pitch
and relative pitch.’ Neurobiology, 95, 3172-3177.