Professional Documents
Culture Documents
In
the
traditional
aesthetic
view
formed
by
the
end
of
the
18th
century,
the
meaning
of
music
originates
from
the
inherent
values
of
the
music
structures
recorded
by
the
composer,
and
illustrated
by
the
performer.
This
view
holds
that
the
listener
finds
a
piece
of
music
meaningful
when
the
performer
“convinces”
him
in
the
authenticity
of
a
certain
emotional
state
inherent
to
the
performed
music.
Once
the
listener
“believes”
the
performer
and
starts
experiencing
the
projected
emotional
state,
he
opens
up
to
the
appreciation
of
other
aspects
of
music.
Disbelief
in
the
performer
and
the
lack
of
emotional
communication,
on
the
other
hand,
blocks
the
capacity
of
the
listener
to
identify
the
changes
in
music
and
therefore
prevents
aesthetic
evaluation
–
no
matter
how
virtuosic
or
grammatically
“correct”
the
music
appears,
the
audience
remains
unmoved.
The
bulk
of
the
music
repertoire
most
often
heard
in
concert
halls
and
studied
in
music
schools
consists
of
performances
created
in
accordance
with
this
view.1
The
philosophy
of
music
education
therefore
heavily
leans
toward
regarding
musical
structures
as
a
primary
source
for
deriving
meaning
and
formulating
interpretation.
These
structures
are
considered
much
more
important
than
the
biographical
circumstances
of
the
composer,
social
environment
during
the
time
of
composition,
the
initial
reception
of
the
music,
the
dramatic
context
associated
with
it,
etc.
-‐
the
consensus
in
music
schools
tends
to
regard
all
these
notions
as
secondary
to
musical
structures
as
the
basis
for
performance
renditions.
The
same
view
is
usually
applied
onto
musical
traditions
other
than
classical
music
today.
They
are
also
seen
to
operate
on
the
premise
of
some
musical
grammar
that
prescribes
which
musical
structures
ought
to
be
used,
and
the
performer
is
expected
to
1
Of
course,
the
connection
of
music
with
emotion
remains
a
matter
of
aesthetic
preference.
Some
people
might
choose
to
create
or
consume
music
that
is
deliberately
unemotional.
Noteworthy
is
the
anti-‐emotional
rhetoric
of
Stravinsky
(Stravinsky
1936,
83).
But
even
more
noteworthy
is
his
double-‐standard:
nowhere
in
his
copious
writings
did
he
ever
name
a
performer
who
would
have
over-‐emotionalized
his
music,
moreover,
he
pronounced
Samuel
Dushkin
to
be
one
of
the
best
interpreters
of
his
works
(Stravinsky
2005,
355–6),
while
Dushkin,
the
pupil
of
Auer
and
Kreisler,
as
expected
from
a
world
class
violinist
in
the
first
half
of
the
20th
century,
played
Stravinsky
with
utmost
Romantic
passion
(as
evident
from
his
recordings).
It
seems
that
emotional
affinity
of
music
tends
to
find
a
way
to
penetrate
even
those
music
works
that
are
claimed
to
be
devoid
of
it.
2
know
that
grammar,
and
use
it
to
generate
an
emotionally
charged
music
composition
by
either
improvising
it
or
varying
some
pre-‐existing
model.
The
exact
way
in
which
musical
structures
convey
meaning
remains
open
to
debate,
with
different
disciplines
proposing
different
explanations.
The
only
uncontested
feature
is
that
musical
expression
is
based
on
emotion.
The
majority
of
musicians
involved
in
the
production
of
music
and
educators
that
work
with
performers
who
support
the
traditional
autonomous
aesthetic
theory
hold
emotional
response
as
the
fundamental
mechanism
of
musical
communication.
The
composer
is
believed
to
conceive
a
certain
emotionally
charged
musical
idea
and
construct
a
plot
akin
to
a
dramatic
play
where
the
progression
of
themes
and
their
development
generate
a
flow
of
musical
events,
each
colored
with
its
own
emotional
tint.
The
job
of
the
performer,
then,
is
to
discover
the
plot
and
transform
it
into
a
drama
similar
to
how
the
play
director
takes
the
script
of
a
play
and
creates
a
show
out
of
it.
This
“theatrical”
model
was
coined
by
Franz
Liszt
as
part
of
the
concept
of
public
recital.
The
music
life
as
we
know
it
today
largely
originates
from
the
experiments
of
Niccolo
Paganini.
Paganini
was
the
first
musician
to
start
performing
in
public
from
memory
in
order
to
produce
the
impression
of
sincerity
and
spontaneous
creativity.
He
considered
act-‐out
to
be
a
constituent
of
technical
execution,
often
bringing
in
theatrical
effects,
even
stage
action,
to
accompany
his
performances.
His
playing
was
designed
as
an
actor’s
monologue,
exposing
the
chain
of
emotions,
characteristic
images,
often
unpredictably
mixed
with
humor,
tragedy
and
passions.
Paganini
forged
the
image
of
a
performing
violinist
as
a
bold
Romantic
hero,
who
challenges
his
fate
in
the
manner
of
Beethoven.2
However,
Paganini’s
love
for
mystery
and
secrecy
in
his
methods
of
producing
his
supernatural
impression
on
audiences
limited
his
impact
on
the
performance
practices
of
his
day.
It
took
Liszt,
with
his
charisma
and
connections
with
conservatoires’
staff
all
over
Europe
to
undertake
a
major
revision
of
the
ways
for
a
performer
to
select
music
for
performance,
work
with
the
score,
build
the
interpretation,
and
present
it
to
the
public.
The
immense
authority
of
a
composer,
performer,
teacher
and
promoter
allowed
Liszt
to
influence
most
of
European
schools
during
the
second
half
of
the
19th
century.
At
the
heart
of
his
innovation
was
the
idea
of
the
necessity
for
a
performance
to
carry
a
poetic
image
representative
of
the
entire
music
composition
being
performed.
Just
as
in
his
own
compositions,
Liszt
could
not
think
musically
without
associations
with
poetry,
literature,
history,
architecture,
etc.
In
the
same
way,
he
could
not
conceive
an
interpretation
of
someone
else’s
work
without
supporting
the
musical
ideas
with
imagery
of
some
kind
–
in
essence
“re-‐composing”
that
work.
This
deeply
personal
attitude,
invoked
by
such
“re-‐composition,”
set
the
tone
for
sincere
and
integral
presentation
of
the
composition
–
a
form
of
sharing
one’s
life
experience
with
the
audience.
2
Paganini
often
hired
an
orchestra
to
play
Beethoven’s
5th
symphony
during
his
violin
recitals
The
practice
of
defining
the
imagery
for
a
piece
of
music
as
the
groundwork
for
its
interpretation
suggests
that
emotion
is
the
primary
component
of
musical
meaning.
If
the
job
of
the
performer
is
to
highlight
for
the
listener
what
is
meaningful
in
the
musical
composition,
and
the
“program
narrative”
is
the
main
content
generated
by
the
performer
(in
addition
to
the
information
obvious
in
the
score),
then
the
emotional
properties
of
the
narrative
manifest
the
meaning
of
the
music.
Psychological
research
supports
this
conclusion:
Hodges
and
Lipscomb
(Lipscomb
and
Hodges
1996)
reviewed
the
studies
on
perception
of
music,
and
were
able
to
identify
seven
types
of
information
retrieved
by
the
listeners
from
auditioning
musical
works
–
all
seven
comprising
an
integrative
knowledge
system
centered
around
emotion.
The
list
consisted
of:
1) Feelings
-‐
the
affective
response
of
the
listener,
sympathetic
to
the
display
by
the
performer
of
features
that
characterize
a
given
emotional
state;
2) Aesthetic
experiences
-‐
the
judgment
of
proportionality,
appropriateness
and
sublimity
in
organization
of
the
expressive
means
within
the
given
composition;
3) Musical
thought
-‐
the
choice
of
the
thematic
material
for
a
music
work
by
its
composer,
and
the
treatment
of
that
material
within
the
composition,
specifically
the
coherence
diversity
in
its
development;
4) Formal
organization
-‐
a
style
of
logical
connection
of
the
musical
elements
together
employed
by
the
composer
as
a
discipline
of
his
musical
thought,
and
grasped,
and
highlighted
by
the
performer;
5) Realization
of
musical
time
-‐
a
unique
experience
of
time
evoked
in
listener
by
the
musical
timing
style
of
the
composition,
reflected
in
relation
of
tempo
to
density
of
musical
events,
correlated
to
the
musical
movement;
6) Personal
identity
-‐
the
set
of
qualities
and
attitudes
that
resonates
with
the
listener’s
self-‐image
and
serves
to
reinforce
or
compensate
for
a
particular
vision
of
himself;
7) Group
identity
-‐
the
set
of
values
that
is
recognized
by
the
listener
as
characteristic
to
a
specific
social
group,
used
to
identify
his
attitude
towards
that
group.
All
seven
types
of
information
conveyed
by
a
music
piece
depend
on
the
emotional
reaction
of
the
listener.
He
experiences
those
feelings
that
have
been
conditioned
in
his
previous
life
to
form
an
emotional
reflex
triggered
by
a
particular
type
of
stimulus.
A
listener
cannot
feel
dignity
in
music,
unless
he
has
developed
the
sense
of
dignity
prior
to
listening.
Aesthetic
evaluation
of
music
can
only
occur
if
the
listener
is
placed
in
an
appropriate
emotional
state.
A
comedy
will
appear
silly
to
a
person
who
is
in
a
serious
mind
set.
A
tragedy
will
appear
ridiculous
to
a
person
in
a
playful
disposition.
Without
identifying
the
target
emotional
condition,
the
listener
will
not
be
able
to
judge
how
harmonious,
appropriate,
or
expressive
the
rendition
of
the
musical
elements
was.
5
A
listener
recognizes
a
particular
musical
thought
in
music
based
on
its
characterization.
It
is
the
change
in
character
that
separates
one
musical
theme
from
another.
So,
in
order
to
follow
the
compositional
plan,
the
listener
has
to
be
stimulated
by
every
change
in
character:
each
instance
of
change
in
the
thematic
material
in
music
ought
to
bring
in
a
new
emotional
state.
In
music,
formal
organization
is
built
around
the
thematic
work
as
well.
One
section
differs
from
another
section
by
utilizing
a
different
theme,
and
more
than
anything
else,
themes
differ
by
their
character.
In
order
to
be
aware
of
the
composer’s
logical
plan
for
the
piece,
a
listener
must
self-‐reflect
on
changes
of
his
feelings
throughout
the
music,
remember
them,
and
relate
them
to
each
other
as
the
composition
progresses.
More
intense
or
diverse
emotional
experience
leads
to
subjective
compression
of
time.
A
shortage
of
emotional
response
to
music
leads
to
boredom
and
the
illusion
of
time
becoming
longer,
or
slower.
The
same
applies
to
the
impression
of
musical
movement.
Emotionally
unmoving
performances
are
not
capable
of
generating
a
sense
of
characteristic
movement
in
music.
Finally,
personal
and
group
identities
are
forged
largely
by
the
experience
of
a
particular
music
work
in
terms
of
its
emotional
content.
Group
identity,
in
particular,
depends
on
the
mechanisms
of
emotional
contagion.
As
we
can
see,
musical
emotion
is
at
the
center
of
the
entire
sphere
of
semantic
connotations
in
music.
Now
lets
ask
the
following
question:
what
exactly
in
the
music
works
as
a
carrier
of
the
emotion,
and
how
is
it
related
to
the
actual
emotion
that
humans
normally
experience?
Obviously,
emotion
cannot
be
encoded
in
a
music
score.
Emotion
is
a
set
of
physiological
changes
in
the
body,
and
cannot
be
“transcoded”
into
notes.
Obviously,
listeners
recognize
particular
combinations
of
sounds
as
possessing
characteristics
similar
to
the
display
of
a
particular
emotion
they
know.
For
instance,
anger
is
associated
with
loud
dynamics,
fast
tempo,
abrupt
articulation,
and
wide
jumps
in
pitch.
Once
the
listener
recognizes
a
similar
set
of
features
in
music,
he
entrains
his
state
correspondingly.
So,
the
mastermind
behind
the
musical
emotion
is
a
performer.
Then,
how
does
a
performer
come
to
grasp
a
particular
emotion
in
music?
Many
musicians
would
agree
that
the
meaning
of
a
work
of
music
is
to
create
a
“virtual
person”
as
a
protagonist
in
some
“virtual
reality”
designated
by
“virtual
time”
(evident
in
rhythm,
meter,
tempo,
articulation,
and
form)
in
conjunction
with
“virtual
space”
-‐
a
subjective
impression
from
musical
movement
(interaction
of
rhythm,
meter,
tempo
and
articulation
with
melody,
harmony
and
texture).
The
adventures
of
such
a
“virtual
person”
are
usually
derived
from
the
development
of
thematic
material
within
the
composition.
What
eventually
happens
with
a
“virtual
person”
in
“virtual
reality”
constitutes
a
sort
of
internal
cinema
-‐
played
back
in
the
listener’s
and
the
performer’s
mind
alongside
the
course
of
the
music.
The
only
difference
from
film
here
is
the
inherent
vagueness
of
6
imaging,
leading
to
substantial
variations
in
renditions
of
the
same
musical
composition
by
different
performers,
as
well
as
variations
in
experiencing
the
same
performance
by
different
listeners.
Watt
and
Ash
(Watt
and
Ash
1998)
conducted
an
experimental
study
and
found
that
listeners
routinely
attribute
traits
such
as
gender,
age,
temper,
and
mood
to
the
music
they
hear.
In
essence,
they
construct
a
“virtual
person”
as
representative
of
the
thematic
material
in
music
in
a
way
similar
to
how
a
reader
constructs
a
protagonist
of
a
story
in
their
mind
while
reading
a
novel.
The
listener
then
starts
interacting
with
the
personality
of
this
“virtual
person”
–
which
can
lead
to
quite
complex
emotional
changes
depending
on
how
the
listener
identifies
himself.
The
vagueness
of
this
musical
personality
and
the
openness
of
the
interaction
of
that
personality
with
the
personality
of
a
listener
are
actually
advantages
of
music
medium
over
film,
which
has
a
precise
and
concrete
representation
of
events.
Watching
the
same
film
over
again
will
not
alter
the
personality
of
a
protagonist
of
that
film.
Listening
to
the
same
music
composition
performed
by
different
performers
can
radically
change
the
“virtual
person”
–
the
more
creative
the
performer,
the
greater
the
differences
in
this
“virtual
person”.
The
greater
variety
of
emotional
experience
from
musical
sources
is
one
of
the
reasons
why
music
became
the
format
of
communication
dedicated
to
emotion.
Exact
repetition
is
known
to
increase
predictability
and
induce
relaxation.
Relaxation
is
an
emotional
marker,
and
will
interfere
with
transmission
of
a
whole
range
of
emotions
(such
as
anger,
anxiety,
jealousy,
etc.).
Variance
in
performance
allows
for
a
fair
representation
of
all
emotional
states
by
keeping
the
senses
of
the
listener
alert,
registering
the
small
changes
and
discrepancies
from
previous
performances.
There
is
ample
evidence
that
many,
if
not
most,
listeners
perceive
music
to
reflect
their
own
personality:
the
way
they
think,
their
likes
and
dislikes,
strengths
and
weaknesses.
One
of
the
leading
preferences
in
choice
of
music
to
listen
to
is
known
to
be
the
need
to
confirm
one’s
own
identity.
Gabrielsson
and
Lindstrom
Wik
(Gabrielsson
and
Wik
2003)
investigated
the
reports
of
“strong
experiences
related
to
music”
(SEM)
by
900
subjects.
Reactions
included
sensation
of
freedom,
inspiration,
upliftedness;
comfort,
hope,
power,
relief
–
obviously,
all
of
which
appear
to
be
dispositions
towards
actions.
Evidently,
music
serves
as
an
agent
capable
of
motivating
people
towards
specific
behaviors
in
a
way
that
is
essentially
similar
to
real
life
encounters.
In
fact,
it
is
likely
that
music
provides
even
more
intense
emotional
action,
since
many
participants
remembered
their
responses
from
long
time
ago
and
kept
it
private
-‐
never
relating
their
SEM
to
anyone
else.
The
advance
of
cognitive
sciences,
and
especially
of
neurophysiology,
during
the
last
twenty
years,
has
provided
new
evidence
in
support
of
the
traditional
view
of
emotions
as
the
basis
of
musical
communication.
For
about
two
thousand
years
emotions
have
been
understood
as
“the
appraisals
of
consciously
detected
feeling
states
named
with
familiar
words”
(Kagan
2007,
1–54).
7
The
consensus
between
scholars
was
that
emotion
was
a
mentalized
form
of
feeling,
common
for
all
individuals,
and
therefore
designated
by
a
group
of
common
words
with
essentially
the
same
meaning
(anger,
rage,
ire,
annoyance).
This
traditional,
“grammarian,”
view
stemmed
from
the
Aristotelian
tradition
of
focusing
on
the
way
in
which
a
concept
is
expressed
in
language.
Modern
neuroscience
examines
emotions
as
they
are,
and
not
as
they
are
named
or
referred
to.
Subsequently,
the
definitions
of
emotions
and
feelings
have
been
redrawn.
The
traditional
meaning
of
“emotion”
is
addressed
in
affective
neuroscience
by
the
term
“feeling”
–
which
represents
a
person’s
subjective
evaluation
and
perception
of
their
emotional
state.
“Emotion”
is
seen
as
a
“largely
automated
program
of
actions
concocted
by
evolution”
-‐
a
stable
mechanism
of
triggering
a
particular
physiological
state
that
is
vitally
important
for
the
survival
of
an
organism
and
a
species.
“Mood”
is
separated
into
its
own
category,
regarded
as
a
continuous
affective
state
that
is
relatively
diffused,
without
an
acute
mental
awareness
of
its
character,
and
detached
from
the
environmental
stimuli
(Damasio
2012,
116).
So,
what
happens
during
an
audition,
is
that
the
listener
constructs
a
virtual
person
of
a
certain
gender,
age,
temper
and
mood,
and
assigns
to
that
person
a
particular
emotional
state.
This
state
keeps
changing
throughout
the
piece
of
music,
prompting
the
listener
to
imagine
some
sort
of
happening,
the
chain
of
events
that
are
likely
to
trigger
a
specific
chain
of
emotional
states.
Because
the
plot
here
is
set
by
the
emotional
descriptors,
it
leaves
a
very
wide
range
for
fantasy:
each
listener
can
envisage
the
story
of
music
in
his
own
way.
If
mood
and
feeling
are,
in
essence,
perceptions
-‐
then
emotion
is
basically
an
action
generated
by
the
organism
in
response
to
selective
types
of
stimuli.
Once
the
senses
capture
data
and
the
brain
recognizes
the
presence
of
stimuli
that
have
been
registered
previously
as
related
to
certain
emotional
responses,
the
brain
activates
a
number
of
emotion-‐triggering
regions
(amygdala
and
frontal
lobe
cortex).
The
endocrine
glands
and
subcortical
nuclei
start
secreting
chemicals
like
cortisol
or
adrenalin.
Injected
into
the
bloodstream,
these
chemicals
quickly
reach
the
internal
organs
and
regulate
their
activity
in
a
way
stereotypical
to
a
given
emotion
(e.g.
surge
of
adrenalin
enables
intense
action,
like
fighting
in
anger).
This
pattern
of
changes
in
the
physiological
state
of
the
body
is
fixed
for
each
emotion.
Though
the
context
of
invocation
for
a
given
emotion
can
be
radically
different
each
time,
all
instances
of
the
experience
of
a
particular
emotion
by
a
given
individual
are
essentially
identical
(ibid.).
It
is
this
structural
stability
of
emotion
that
makes
it
possible
for
music
to
function
as
a
language
of
emotions.
Contrary
to
public
opinion,
the
meaning
of
music
is
no
less
clear
than
the
meaning
of
speech.
Just
as
the
way
in
which
the
word
“chair”
refers
to
something
on
which
a
person
can
sit
comfortably,
the
pattern
of
pitch,
rhythm
and
harmony
in
music
refers
to
a
specific
physiological
state.
Stable
structure
of
emotion
is
connected
to
stable
structure
of
sounds
–
8
what
can
be
called
a
“musical
idiom”:
an
audible
pattern
of
sounds
segregated
into
a
group
that
would
be
associated
with
a
particular
mood,
attitude,
character,
or
idea
by
different
listeners
upon
numerous
auditions.
In
linguistics,
an
idiom
is
a
group
of
words
commonly
used
together,
where
the
meaning
of
the
group
is
not
equal
to
the
sum
of
meanings
for
the
constituent
words
(as
in
“pulling
one’s
leg”).
In
order
for
the
idiom
to
be
understood
properly,
the
listener
has
to
know
the
convention
(Saeed
2015,
59–60).
In
music,
the
word
“idiom”
is
used
to
mean
“a
characteristic
trait
of
the
specific
progression
of
tones
-‐
a
trait
by
which
the
listener
can
identify
that
the
music
belongs
to
a
certain
type”.
For
instance,
a
melody
based
on
the
sounds
of
a
triad,
with
an
overall
ascending
direction,
in
a
major
key,
and
loud
dynamics
is
usually
recognized
as
“fanfare”
and
regarded
as
idiomatic
to
the
tone
of
the
trumpet
(i.e.
Tchaikovsky
–
the
opening
of
Italian
Capriccio).
Similarity
to
the
linguistic
idiom
is
obvious:
the
listener
knows
how
the
sounds
are
made
on
the
trumpet
and
what
kinds
of
melodic
patterns
are
the
easiest
to
produce
on
it.
Then
the
stereotypical
properties
of
trumpet
music
become
projected
on
the
musical
idiom
of
fanfare:
trumpet
(and
other
brass
instruments)
calls
have
been
used
in
Western
culture
as
signals
to
announce
something
important
for
the
public
(Huron
and
Berec
2009).
Therefore,
the
idiom
of
fanfare
connects
to
emotion
of
courage.
The
linkage
goes
something
like
this:
public
announcements
usually
are
related
to
public
authorities,
authority
has
to
do
with
power,
and
display
of
power
irrespective
of
circumstances,
even
in
situation
of
danger,
characterizes
courage.
So,
listening
to
the
opening
of
Italian
Capriccio
by
Tchaikovsky
makes
a
listener
who
is
aware
of
the
significance
of
brass
music
in
Western
civilization
experience
the
physiological
condition
that
is
essentially
the
same
as
what
one
feels
when
he
stands
up
for
his
beliefs
regardless
of
the
repercussions.
It
may
appear
that
the
knowledge
of
the
convention
for
the
fanfare
is
something
esoteric,
which
would
make
emotional
communication
through
music
way
too
exclusive
for
music
to
be
considered
a
“language.”
But
competence
in
language
is
much
wider
than
expertise
in
language.
Most
people
are
passively
aware
of
things
they
cannot
actively
express
in
words.
In
the
same
way,
familiarity
with
fanfare
does
not
require
explicit
knowledge
of
the
history
of
music,
nor
the
exact
structures
that
comprise
it.
Most
people
who
lived
for
few
years
in
a
Western
country
would
have
witnessed
ceremonies
that
involve
fanfares:
national
celebrations,
military
funerals,
openings
of
important
sport
events,
etc.
Subsequently
recognizing
similar
sound
structures
in
the
context
of
a
soundtrack
in
some
movie
would
reinforce
its
musical
meaning,
and
enable
one
to
identify
fanfare
within
the
context
of
a
musical
work,
like
Capriccio
Italienne.
9
The
capacity
for
intuitive
competence
in
musical
idioms
is
not
unique
to
the
Western
classical
music.
Any
naturally
evolved
music
culture 3
operates
on
the
same
semiotic
principles:
there
is
a
pool
of
patterns
of
melody,
rhythm,
meter,
articulation,
and
texture,
which
are
reused
in
every
musical
composition
and
require
from
the
music
user
the
knowledge
of
conventional
reference
to
corresponding
emotional
conditions.
User
of
any
kind
of
traditional
music4
is
usually
competent
in
the
music
of
his
native
culture.
The
configuration
of
the
musical
aspects
of
pitch,
rhythm,
and
harmony
determines
the
structure
of
a
music
idiom.
The
configuration
of
features
of
a
physiological
state
constitutes
the
structure
of
emotion.
A
cultural
convention
stitches
the
structure
of
a
music
idiom
to
the
structure
of
an
emotion
–
thus
providing
the
foundation
for
a
music
“language”.
A
person
with
sufficient
musical
training
has
enough
command
of
music
idioms
to
conceive
them
mentally
–
to
think
in
terms
of
music.
Laypeople
cannot
rely
on
musical
thinking.
Their
command
of
music
is
more
passive
than
active.
This
might
suffice
for
simple
forms
of
music
making,
such
as
folk
music.
But
for
art
music,
such
as
classical
music
or
jazz,
their
lack
of
ear
training
will
be
an
obstacle
in
the
detection
of
music
idioms
in
a
musical
composition,
whether
improvising
music
or
playing
by
ear.
They
need
an
emotional
agent
in
order
to
convert
the
emotional
content
of
a
music
language
into
an
artwork.
They
need
a
master
performer
who
is
capable
of
converting
the
musical
structures
into
emotions.
Then
the
layperson
will
sympathize
with
the
performance
and
generate
the
corresponding
emotional
states.
To
bridge
the
gap
between
musical
idiom
and
human
emotion,
we
must
therefore
consider
another
structural
entity
–
the
musical
emotion.
A
musical
emotion
is
the
emotion
implied
by
the
notes
of
a
score,
encoded
in
the
musical
text
by
means
of
cultural
convention,
and
designed
to
be
retrieved
and
re-‐enacted
by
a
performer.
The
structure
of
this
emotion
is
very
complex
and
the
rules
for
interpretation
of
these
emotions
are
vague,
strongly
intuitive,
and
are
intermingled
with
numerous
exceptions
and
conditions.
That
is
3
Artificially
invented
music
systems,
such
as
dodecaphony
by
Schoenberg
and
Webern,
or
serial
music
by
Boulez
and
Stockhausen,
are
the
exception.
They
denounce
the
pre-‐existing
grammar
and
lexicon
of
musical
expressions,
and
substitute
them
with
completely
new
grammar
and
lexicon.
This
radical
approach
makes
intuitive
competence
impossible
in
relation
to
these
types
of
music,
as
well
as
in
later
compositional
techniques
that
followed
the
same
trend
(i.e.
new
complexity).
4
A
number
of
genres/styles
of
Western
popular
music
constitute
a
special
case.
Their
use
of
idioms
usually
becomes
severely
limited
by
the
commercial
goals
of
production:
even
where
a
style
of
popular
music
originates
from
some
folk
prototype,
the
popular
version
usually
drops
many
idioms
and
simplify
the
rest
of
the
idioms
in
order
to
widen
the
base
of
potential
buyers.
The
prime
goal
of
such
music
is
for
it
to
be
liked
by
as
many
people
as
possible,
without
requiring
them
to
learn
any
idioms.
This
usually
makes
the
mass-‐produced
Western
popular
music
rely
on
the
superficial
“phonic”
characteristics
of
sound
as
well
as
non-‐musical
means
of
communication
(lyrics,
titles,
histrionic
aspects
of
presentation,
costumes,
stage
design,
etc.)
–
rather
than
musical
idioms
–
in
emotional
communication.
10
why
it
takes
years
for
a
person
to
learn
music
interpretation,
and
there
are
relatively
few
masters
who
can
do
it
well.
Musical
emotion
has
nothing
to
do
with
feeling
(in
the
scientific
sense
of
the
term).
Feeling
is
the
“face”
of
the
emotion
for
the
listener.
One
of
the
leading
specialists
on
emotion,
Antonio
Damasio,
defines
feeling
as
a
perception
of
an
emotional
state;
the
mental
awareness
of
the
physiological
state
of
the
body
as
well
as
the
concurrent
state
of
mind
(Damasio
2012,
117).
In
relation
to
music
listening,
an
example
would
be
the
experience
of
thrills
(goose-‐bumps
from
strong
emotional
reaction)
–
the
listener
becomes
aware
of
them
and
connects
them
to
previous
instances
of
feeling
in
the
same
way.
In
this
way,
causes
and
behavioral
reactions
from
the
past
connect
with
the
feeling
of
“now,”
and
the
feeling
of
thrill
becomes
structured
into
the
emotion
of
pleasure.
Together,
feeling
and
emotion
go
hand
in
hand,
forming
a
cycle:
perception
of
data
triggers
an
emotional
reaction
that
starts
in
the
brain,
involving
glands,
which
spreads
over
the
entire
body.
At
this
point,
the
feeling
stage
kicks
in
-‐
activating
the
new
brain
regions
and
concluding
the
cycle,
as
the
brain
becomes
conscious
of
the
overall
state
of
the
body
and
relates
it
to
a
particular
mode
of
behavior.
The
principal
difference
between
emotion
and
feeling
is
that
feeling
is
rather
free-‐
floating
and
even
optional
-‐
it
can
follow
the
emotional
response,
or
be
altogether
omitted,
depending
on
the
attention,
urgency
and
self-‐awareness
of
a
person.
Emotion,
on
the
other
hand,
is
highly
structured
and
automated.
Modern
psychology
defines
emotions
as
bio-‐
behavioral
systems
that
consist
of
at
least
four
core
components
(Humrichouse
et
al.
2007):
(a) a
subjective
experience
(e.g.
seeing
a
dog
that
is
perceived
as
“dangerous”);
(b) a
physiological
reaction
(e.g.
fear,
manifested
by
increased
heart
rate
and
general
sympathetic
activation);
(c) overt
expression
of
the
physiological
state
(raised
eyebrows
and
wide-‐open
eyes);
(d) a
behavioral
response
(a
chosen
strategy
to
cope
with
the
situation,
such
as
a
decision
to
freeze
or
flee).
These
components
follow
each
other
in
a
coordinated
manner
in
a
short
period
of
time,
in
the
order
of
seconds,
mostly
beyond
the
control
of
an
individual.
Only
the
behavioral
response
stage
is
partially
controllable,
when
the
mind
has
the
capacity
to
moderate
or
modify
the
behavior
rather
than
acting
upon
the
first
impulse.
In
relation
to
musical
emotion,
this
fourth
component
is
usually
suppressed:
listeners
do
not
run
away
from
the
concert
hall
upon
hearing
scary
music;
however
they
can
root
or
cheer
while
listening
to
some
fiery
march.
The
exact
choice
for
behavior
depends
on
the
place
and
purpose
of
application
of
music.
The
time
frame
for
musical
emotion
to
kick
in
is
shown
to
be
about
8
seconds.
A
dedicated
study
of
emotional
response-‐time
to
music
found
that
music
expertise
has
little
to
do
with
the
speed
of
emotional
response
in
listeners,
and
is
significantly
accelerated
in
cases
where
the
music
is
already
familiar
to
the
listener
(Bachorik
et
al.
2009).
11
Emotions
Across
Cultures
The
fact
that
the
display
patterns
for
the
same
emotion
stay
very
similar
across
different
species
testifies
to
the
importance
of
emotion
to
the
survival
in
the
animal
kingdom.
Thus,
the
baring
of
teeth,
characteristic
in
display
of
anger,
was
adaptive
as
a
preparation
for
biting
in
attack;
narrowing
of
the
eyes
was
protecting
the
eye
from
injury
in
combat.
High
reactivity
towards
specific
stimuli
early
in
life
also
points
to
the
evolutionary
origin
of
emotions:
goslings
show
fear
when
exposed
to
the
shape
resembling
a
hawk
without
ever
seeing
an
actual
hawk.
Evolutionary
factors
are
found
to
speed
up
linking
a
given
environmental
stimulus
with
emotion,
of
enduring
such
a
connection,
and
mediating
the
intensity
of
the
emotion.
It
appears
that
emotions
are
somewhat
pre-‐programmed
for
their
rapid
establishment,
differentiation,
and
fine-‐tuning
in
the
best
interests
of
an
individual.
A
number
of
neurobiologists,
like
Paul
MacLean
or
Jaak
Panksepp,
speak
about
emotional
circuits
embedded
subcortically
within
the
mammalian
brain.
Such
a
genetically
dictated
emotional
operating
system
allows
newborn
animals
to
begin
responding
adequately
to
their
environment
at
the
stage
when
these
animals
lack
any
cognitive
means
of
realization.
The
circuits
responsible
for
dreaming,
anticipation,
the
pleasures
of
eating
as
well
as
the
13
consumption
of
other
resources,
anger,
fear,
love,
lust,
maternal
acceptance,
grief,
play,
and
joy
ensure
effective
start
in
the
process
of
mental
development
(Panksepp
1998,
4).
Panksepp
explains
the
organization
of
these
emotional
circuits
in
terms
of
computer
hardware.
The
brain-‐bio-‐computer
has
an
operating
system
installed
in
the
read-‐only
memory
(ROM)
from
the
moment
of
birth.
ROM
functions
are
genetic
and
reside
in
lower,
subcortical
regions
of
the
brain.
More
intricate
functions
of
learning
and
cognition
are
elaborated
in
the
random-‐access
memory
(RAM),
which
is
more
flexible,
faster
and
can
be
programmed
by
the
user
of
the
computer.
These
functions
are
evolutionarily
more
recent
and
are
concentrated
in
the
neo-‐cortex.
All
mammals
appear
to
have
very
similar
ROM
functions,
while
humans
rely
on
far
more
developed
RAM-‐like
neuro-‐computational
space
than
other
mammals.
The
basic
emotions
belong
to
ROM,
whereas
cognition
operates
in
RAM
(ibid.,
20).
By
this
scheme,
complex
emotions
and
adjusted
basic
emotions
(such
as
sense
of
fear
moderated
in
certain
situations,
i.e.
a
stage
fright)
are
entered
in
RAM,
as
cognition
keeps
regulating
the
pre-‐programmed
emotional
responses
to
the
needs
of
an
individual.
The
capacity
for
fine-‐tuning
emotions
is
crucial
for
success,
especially
in
the
modern
urban
environment,
where
complex
social
organization
makes
an
individual
cater
to
multiple
contacts,
services,
and
tasks.
For
one
situation,
it
might
be
beneficial
to
have
a
heightened
sense
of
fear
(investment
in
the
stock
market),
whereas
for
another
-‐
to
have
fear
minimized
(stage
fright).
Failure
to
optimize
an
emotion
is
directly
connected
to
negative
outcomes
that
would
hurt
the
career
of
that
individual.
Music,
with
its
capacity
to
continually
exercise
emotional
response,
and
with
the
individual’s
control
over
the
choice
of
a
particular
kind
of
music,
provides
a
tool
par
excellence
for
calibrating
emotions.
The
fixed
structure
of
emotion,
per
se,
as
well
as
the
fixed
patterns
in
the
expression
of
emotional
display
(frowning,
smiling),
testify
to
the
objectivity
of
emotions,
thereby,
correcting
the
older
scientific
views
dominant
in
the
20th
century
which
held
emotion
to
be
as
something
inherently
changeable
and
subjective.
If
the
mechanism
of
basic
emotions
is
genetically
embedded,
and
their
communication
is
universal,
as
indicated
by
the
most
recent
research
that
distinguishes
between
universality
of
basic
and
non-‐basic
emotions
(Matsumoto
and
Hwang
2012),
then,
it
becomes
possible
to
claim
that
people
experience
each
of
the
basic
emotions
in
essentially
the
same
way.
Given
this
objective,
the
structured
status
of
emotion,
basic
musical
emotions
become
a
likely
candidate
for
the
foundation
of
a
universal
“emotional
language”
of
music,
provided
that
the
musical
descriptors
of
musical
emotions
(pitch,
rhythm,
meter,
harmony,
texture)
do
not
require
any
conventional
referents
(knowledge
of
a
specific
configuration
of
pitch/rhythm/meter/harmony/texture
pattern).
Many
elements
of
musical
composition
can
be
emotionally
interpreted
on
this
basis,
and
are
clear
in
their
meaning
purely
by
the
virtue
of
their
physical
properties:
loudness
–
standing
for
intensity,
register
–
for
size
(K.
K.
Evans
and
Treisman
2011),
tempo
–
for
motion
(Feldman,
Epstein,
and
Richards
1992).
There
are
a
number
of
acoustic
attributes
of
sound
that
do
not
require
knowledge
of
convention
to
be
understood
-‐
they
act
on
a
synesthetic
basis,
by
automatically
converting
an
auditory
stimuli
into
some
sensory
images
or
qualities
of
other
sensory
modalities,
such
14
as
vision
or
palpation
(Marks
2014).
Such
cross-‐modal
integration
seems
to
be
a
widespread
phenomenon,
presenting
a
norm
rather
than
anomaly
for
humans
(Parise
and
Spence
2009).
It
is
just
that
some
people
are
conscious
of
synesthetic
connections,
whereas
others
are
not
(Simner
2012).
The
ramifications
of
this
for
emotional
communication
are
far-‐reaching.
An
emotional
reaction
can
be
primed
by
a
visual
stimulus,
and
transferred
onto
the
auditory
domain
by
means
of
shared
neural
resources.
There
is
experimental
support
for
such
cross-‐modal
interaction
of
auditory
and
visual
emotions
for,
both,
simultaneous
stimuli
and
sequential
processing
(Logeswaran
and
Bhattacharya
2009).
An
example
of
such
interaction
could
be
found
in
the
cross-‐modal
correspondence
between
the
tonality/mode
of
music
and
colors
–
evidently,
mediated
by
emotion
(Palmer
et
al.
2013).
In
practice,
emotional
reference
works
in
a
complex
manner,
integrating
a
specific
configuration
of
auditory
attributes:
thus,
music-‐color
associations
involve
a
combination
of
tempo,
mode,
timbre,
and
harmony
(Tsang
and
Schloss
2012).
However,
some
important
auditory
attributes
have
demonstrated
stable
synesthetic
associations
that
can
bear
definitive
emotional
values.
• The
connection
between
loudness
and
intensity
(K.
K.
Evans
and
Treisman
2011)
easily
directs
one
to
perceive
very
soft
sounds
as
tender,
whereas
very
loud
-‐
as
angry.
• The
register/size
correspondence
(Marks
1978,
53)
suggests
cuteness
of
sounds
that
are
very
high
in
pitch,
and
potential
threat
associated
with
very
low
pitches,
since
they
characterize
very
big
creatures
that
can
inadvertently
harm
a
person.
• Similar
correspondence
of
thickness
and
pitch,
discovered
in
infants
(Dolscheid
et
al.
2014),
supports
the
association
of
chords
with
security,
whereas
single
sounds
with
vulnerability.
• Synesthetic
mapping
between
lightness
and
melodic
interval
(Hubbard
1996)
allows
to
associate
a
stepwise
melody
in
high
register
with
light
and
kindness,
versus
leaps
in
low
register
with
darkness
and
evil.
• The
cross-‐modal
link
between
audition
and
tactile
sensation
(Kim
and
Zatorre
2010)
enables
the
vibrotactile
discrimination
of
musical
timbre
and
texture
(Russo,
Ammirante,
and
Fels
2012):
i.e.
pleasant
“velvet”
sound
of
tenor
saxophone
(Nykänen
2004).
There
is
evidence
of
connection
between
the
growling
timbre
of
distorted
electric
guitar
and
the
state
of
aggression
(Tsai
et
al.
2010).
• Musical
articulation
also
carries
cross-‐modal
functionality:
staccato
manner
of
performance
is
associated
with
angular
shapes
and
discomfort,
whereas
legato
articulation,
with
rounded
shapes
and
pleasure
(Weinstein
and
Gridley
2010).
Hearing
of
musical
patterns
with
salient
features
of
register,
loudness,
textural
thickness,
articulation,
melodic
intervallic
size,
and
timbre
–
coordinated
with
tempo
and
metro-‐rhythmic
features,
are
capable
of
invoking
imagery
of
physical
motion
of
very
specific
type,
i.e.
gesture
or
gait
that
can
be
highly
suggestive
of
a
particular
expression
(Clarke
2001).
Thus,
music
can
project
the
impression
of
panting
while
running.
or
of
laziness
while
stretching.
The
connection
between
a
particular
configuration
of
auditory
markers
is
shown
to
provide
a
stable
reference
to
a
certain
cross-‐modal
analog
in
physical
15
reality:
i.e.
descending
melodic
motion
is
associated
with
diminuendo
and
relaxation
(Eitan
and
Granot
2006).
Such
emotional
associations
do
not
really
depend
that
much
on
the
listener’s
familiarity
with
the
conventions
of
musical
grammar
adopted
within
his
music
community.
Unless
his
musical
culture
deliberately
breaks
away
from
the
“natural”
synesthetic
connections
(which
is
not
that
common),
any
listener
has
a
good
chance
of
figuring
out
the
emotional
“meaning”
of
a
music
work
by
just
resorting
to
its
cross-‐modal
functionality.
Universality
of
basic
musical
emotions
across
different
cultures
must
owe
itself
to
such
synesthetic
connections.
There
is
enough
evidence
to
conclude
safely
that
the
basic
emotions
are
universally
recognized
within
the
same
culture:
where
the
presented
music
is
native
to
the
listeners.
In
a
major
experimental
study
(Mohn,
Argstatter,
and
Wilker
2011)
a
group
of
115
university
students
were
able
to
identify
six
basic
universal
emotions
in
completely
unknown
musical
stimuli,
in
most
cases
well
above
chance
level.
The
ability
to
detect
musical
emotions
did
not
seem
to
be
influenced
by
musical
experience
or
personality
traits,
with
recognition
of
happiness
and
sadness
with
extremely
high
reliability
–
testifying
to
the
centrality
of
these
two
emotions
for
Western
music.
Evidently,
laypeople
are
competent
enough
to
recognize
and
match
the
structures
of
musical
idioms
to
the
structure
of
basic
emotions
–
following
the
convention
accepted
within
their
native
culture.
The
research
in
recognition
of
emotions
in
music
across
different
cultures
is
limited,
but
seems
to
point
in
the
direction
of
music
being,
not
only
a
refined
product
of
human
culture,
but
part
of
human
nature
–
a
part
of
humanity’s
genetic
foundation.
Indeed,
those
few
studies
that
examined
recognition
of
musical
emotions
by
the
listeners
across
Western,
Japanese,
Hindustani,
and
Chinese
cultures
reported
that
subjects
did
succeed
in
distinguishing
joy,
sadness
and
anger
in
music
unfamiliar
to
their
cultures.
Early
ability
to
extract
emotion
intentions
from
music,
when
nine
months
old
infants
are
shown
to
discriminate
between
happy
and
sad
music,
also
supports
the
idea
of
biological
preparedness
for
music
(Peretz
2013).
One
of
such
studies
examined
emotional
recognition
of
Western
music
by
German
listeners
and
Mafa
people,
with
music
composed
specifically
for
this
experiment
(Fritz,
Sammler,
and
Koelsch
2006).
The
Mafa
are
one
of
approximately
250
ethnic
groups
that
make
up
the
population
of
Cameroun.
They
are
culturally
isolated
and
live
in
the
Extreme
North
Province
in
the
Mandara
mountain
range.
The
remote
Mafa
settlements
do
not
even
have
electrical
supply,
and
are
inhabited
by
individuals
who
pursue
a
traditional
lifestyle
and
have
not
been
exposed
to
Western
music.
Both,
German
and
Mafa
listeners
identified
all
three
emotions
in
the
music
samples
at
levels
that
were
above
chance
performance.
Among
German
listeners,
recognition
of
happy
music
(99%)
was
slightly
better
than
sad
music
(93%)
or
scary
music
(81%).
Similarly,
among
the
Mafa
listeners,
recognition
of
happy
music
(65%)
was
better
than
sad
music
(49%)
or
scary
music
(48%).
Interestingly,
the
capacity
to
decode
emotional
expressions
was
significantly
correlated
with
the
degree
of
appreciation
of
the
compositions,
supporting
the
studies
on
the
contribution
of
empathy
to
emotional
reaction
to
music.
16
The
universality
of
basic
musical
emotions
could
originate
in
common
neural
pathways
shared
by
emotional
vocalizations
in
verbal
language
and
music.
A
recent
study
(Bowling
et
al.
2012)
examined
the
spectral
analysis
of
happy
and
sad
music
of
Western
(classical)
and
Eastern
(South
Indian
Carnatic)
traditions
versus
the
happy
and
sad
speech
of
American
English
and
Indian
Tamil
speakers.
The
results
demonstrated
close
relationships
between
tonal
variations
in
the
prosody
of
both
verbal
languages
and
both
types
of
music.
Melodic
intervals
in
both
traditions
were
found
to
be
generally
larger
in
happy
melodies,
and
smaller
in
sad
ones.
Similarly,
in
both
languages,
negative/subdued
expressions
were
characterized
by
relatively
small
prosodic
intervals,
whereas
positive/excited
expressions
were
characterized
by
relatively
large
prosodic
intervals.
This
finding
supports
the
previous
reports
of
similarity
between
tempo,
intensity
and
timbre
between
verbal
and
musical
expressions
of
the
same
emotion
(Juslin
and
Laukka
2003).
For
example,
in
both
vocal
and
musical
expressions,
happiness
and
anger
are
associated
with
increases
in
speech
rate/tempo
and
voice
intensity/sound
level,
whereas
sadness
and
tenderness
are
associated
with
decreases
in
speech
rate/tempo
and
voice
intensity/sound
level.
A
review
of
77
studies
of
vocal
expression
and
35
music
performances
seems
to
support
the
conclusion
that
musicians
spontaneously
structure
their
performance
in
a
way
that
resembles
the
overt
expression
of
emotions
in
speech.
The
most
recent
studies
support
this
assumption:
thus,
American
listeners
were
quite
effective
in
identifying
3
basic
emotions
(happiness,
sadness,
&
anger)
in
Korean
folk
songs
(Kwoun
2009).
Yet
another
possibility
is
that
music
and
movement
share
a
common
structure
that
affords
equivalent
and
universal
emotional
expressions.
The
computer
generations
of
matching
patterns
of
music
and
motion
appear
to
trigger
similar
emotional
response
in
American
college
students
and
inhabitants
of
a
Kreung
village
in
Cambodia
(Sievers
et
al.
2013).
The
Kreung
participants’
idea
of
“sad”
was
more
similar
to
the
American
“sad”
prototype
than
to
any
other
emotional
prototype,
and
this
cross-‐cultural
similarity
was
observed
for
all
emotions
except
“angry.”
This
pattern
also
held
for
the
movement
results
when
considered
separately
from
music.
When
the
music
results
were
evaluated
alone,
three
of
the
five
emotions
(happy,
sad,
and
scared)
were
closer
to
the
matching
American
prototype
than
any
nonmatching
prototypes.
These
results
suggest
that
the
dynamic
features
of
emotional
expression
are
cross-‐culturally
universal
for
“angry,”
“happy,”
“peaceful,”
“sad,”
and
“scared”
expression,
and
share
similar
dynamic
contours
in
both
music
and
movement.
The
cross-‐modal
nature
of
musical
expression
must
be
responsible
for
the
ability
of
human
newborns
to
readily
respond
to
emotional
communication
–
together
with
the
genetic
component
of
emotional
receptivity.
In
newborns,
display
of
emotional
organization
can
be
observed
early
in
life,
such
as
disturbed
limb
movement
that
accompanies
the
wiping
of
the
baby’s
nose.
Differentiation
between
emotions
occurs
within
the
first
year
of
life:
distress
at
four
weeks
of
age,
joy
at
six
weeks,
anger
at
four
months,
disgust
-‐
at
six,
and
fear
-‐
at
eight
months
(Witherington
et
al.
2010).
The
development
of
complex
emotions
goes
by
the
same
accelerated
pace.
The
basic
emotions,
universally,
and
likely
genetically
shared
by
all
mammals,
are
believed
to
interact
in
their
mechanisms
and
produce
more
complex,
at
first
secondary,
and
then
tertiary
emotions.
As
the
infant
grows,
so
does
his
awareness
of
his
surroundings.
He
becomes
17
involved
in
more
complex
interactions
with
his
caretakers.
Researchers
report
observing
such
complex
emotions
as
shyness
in
four
month
olds,
jealousy
in
babies
of
about
8
months,
and
embarrassment
in
the
ages
between
nine
months
and
two
years
(Draghi-‐
Lorenz,
Reddy,
and
Costall
2001).
The
structural
consistency
of
the
primary
emotions
is
what
allows
more
complex
emotions
to
form.
In
order
for
a
complex
emotion
to
evolve,
each
of
the
emotional
components
that
comprise
that
new
state,
important
in
the
life
of
that
given
individual,
has
to
be
continuously
re-‐experienced
until
the
new
state
is
sufficiently
reinforced
to
be
remembered
by
the
organism.
Despite
their
non-‐genetic
component,
it
is
likely
that
complex
emotions
are
also
transferrable
across
cultural
barriers.
Thomson
and
Balkwill
(Thompson
and
Balkwill
2010)
define
a
special
category
of
markers
responsible
for
emotional
connotations
across
cultures
-‐
psychophysical
cues
that
include
rate,
intensity,
timbre
and
pitch,
and
along
with
culture-‐specific
cues
(e.g.
harmonic,
rhythmic
patterns)
enhance
the
recognition
process
for
members
of
the
same
culture.
These
psychophysical
cues
are
not
merely
the
result
of
innate
processes,
but
also
reflect
other
systemic
constraints:
i.e.
musical
expression
of
high-‐arousal
emotions
(like
anger
or
excitement)
may
be
characterized
by
attributes
that
reflect
the
increased
oxygen
requirements
associated
with
high-‐arousal
states
(faster
respiratory
rate).
Because
the
connection
between
oxygen
requirements
and
arousal
states
is
ubiquitous,
certain
attributes
of
high-‐arousal
music
are
likely
to
occur
across
cultures.
When
cultural-‐specific
cues
are
unfamiliar
or
absent,
listeners
may
still
perceive
psychophysical
cues
such
as
faster
tempo
and
increased
[auditory]
volume.
These
cues
provide
listeners
with
a
general
understanding
of
the
intended
emotion
even
for
unfamiliar
musical
styles
–
providing
a
work-‐around
when
the
listener
has
no
knowledge
of
that
culture’s
musical
conventions.
The
connection
between
music
and
emotion
is
quite
well
established:
79
experimental
studies
between
1892
and
2000
have
found
a
stable
connection
between
the
recognition
of
at
least
a
few
emotional
states
or
emotional
conditions
and
specific
pieces
of
music.
The
accuracy
varied
depending
on
the
emotion
and
music.
Accuracy
was
very
high
for
gaiety,
joy,
and
assertion;
lower,
but
still
well
above
chance
level
for
calm,
sorrow,
tenderness,
and
yearning;
but
less
direct
with
the
latter
ones,
which
were
sometimes
confused
with
each
other.
The
ability
to
discriminate
musical
emotional
expressions
had
no
relation
to
general
intelligence
or
musical
training,
and
carried
only
low
correlations
with
scores
on
the
Seashore
Measures
of
Musical
Talents,
suggesting
that
emotional
communication
through
music
is
a
general
function
for
any
human
being.
Overall,
recognition
stays
reliable
for
positive
emotions
of
gaiety,
happiness,
joy,
mischievousness;
negative
emotions
of
sadness,
sorrow;
anger-‐hate;
dullness;
states
of
high
arousal
and
low
arousal
(excitement/calm)
and
social
emotions
of
tenderness,
and
dignity
(Gabrielsson
and
Juslin
2003).
Emotional
communication
through
music
appears
to
operate
on
objective
grounds,
at
least
with
an
adult
audience
that
shares
the
same
cultural
background
as
the
creator
of
the
music.
A
musical
emotion
is
realized
by
the
musician,
brought
to
life
through
performance,
identified
by
the
adult
listener,
who
then
generates
the
same
emotional
state
via
the
mechanisms
of
emotional
contagion.
Such
a
chain
supports
the
objective
perspective
of
18
musical
emotion.
This
is
not
to
say
that
there
is
no
subjective
component
in
it.
Like
verbal
expression,
musical
expression
involves
both
facets:
the
information
conveyed
through
the
act
of
perception
and
auditory
processing
of
the
signal
is
objective,
but
the
individual
listener’s
associations
and
experience
of
the
physiological
response
is
subjective.
The
objective
basis
of
emotion
lies
in
the
shared
physiological
structures
of
the
basic,
primary
emotions.
In
linguistic
terms,
this
objective
aspect
can
be
called
“denotation”.
Musical
emotion
denotes
a
certain
set
of
physiological
conditions
in
listeners.
The
subjective
aspect
of
emotion
can
be
called
“connotation”:
musical
emotion
invokes
not
only
a
sympathetic
response
in
listeners,
but
it
also
triggers
certain
attitudes,
evaluations
and
resolutions
from
the
listener’s
past
experience,
unique
to
that
individual.
For
instance,
a
sad
expression
can
connote
nostalgia
in
one
person,
melancholy
in
another,
and
misanthropy
in
a
third.
Such
variance
depends
on
the
post-‐emotional
mentalization
at
the
moment
of
listening,
combined
with
the
person’s
memory
of
the
past
experience.
The
discrepancy
between
connotational
and
denotational
emotional
“meanings”
is
responsible
for
apparent
contradictions
in
semantic
values
of
the
very
same
music
in
eyes
of
different
people.
Using
an
example
of
Stephan
Vitas,
many
people
hear
the
opening
of
Beethoven’s
9th
Symphony
as
“man
meets
the
Creator
in
the
great
hall
of
the
universe”;
yet
on
a
radio
talk
show
one
woman
said
she
heard
the
same
passage
as
“the
rapist
confronts
his
victim.”
It
appears
that
both
emotions
deny
each
other,
therefore
testifying
towards
arbitrariness
of
musical
emotion.
However,
the
contradiction
in
this
example
is
illusory.
What
matters
here
is
not
the
conceptual
conclusion,
but
the
perception
prior
to
it.
Musical
communication
handles
matters
up
to
this
point:
both,
an
admirer
of
Beethoven,
and
this
woman
agree
in
recognizing
the
strength
and
awe
represented
by
Beethoven’s
configuration
of
pitch,
meter,
rhythm
and
texture.
Beyond
that
comes
the
turn
for
cognition
of
feeling
and
psychological
interpretations.
But
this
is
no
different
than
reaction
to
speech,
where
the
realization
of
being
offended
often
follows
after
some
deliberation
of
what
was
heard.
If
no
one
would
dispute
the
capacity
of
speech
to
convey
emotion,
then
the
same
should
be
reserved
for
music.
Emotional
communication
through
music
operates
primarily
through
emotional
denotation,
leaving
emotional
connotation
as
an
individualized
response
of
each
listener.
Musical
experience
is
shown
to
be
very
private.
Experimental
investigation
of
psychological
and
physiological
reactions
(skin
conductance
response)
while
listening
in
a
social
setting
versus
private
listening
demonstrates
(H.
Egermann
et
al.
2011),
beyond
any
doubt,
that
listening
to
music
privately
is
more
arousing
than
listening
in
a
group.
There
are
more
chills
reported
during
solitary
listening,
and
the
intensity
of
excitement
is
substantially
higher.
This
leads
to
believe
that
emotional
connotation
is
the
primary
focus
of
emotional
response
to
music
for
an
individual.
Emotional
denotation,
then,
is
needed
for
social
reasons
–
to
keep
an
individual
in
tune
with
the
social
group
and
cultural
environment.
It
is
important
to
remember
that
language
of
music
has
been
shaped
by
the
emotional
denotation,
and
emotional
communication
would
not
be
possible
without
it.
Emotional
19
denotation
constitutes
the
principal
meaning
of
music,
and
emotional
connotation
builds
upon
it.5
Karl
Pribram
was
one
of
the
first
neuroscientists
to
put
forward
a
model
of
emotional
semiosis
for
music
–
an
explanation
of
how
denotative
emotional
meaning
is
formed
(Pribram
1982).
According
to
his
model,
a
particular
music
structure
obtains
its
meaning
as
the
listener
builds
a
solid
memory
of
how
he
feels
while
listening
to
that
structure
–
the
listener
averages
numerous
instances
of
listening
to
it
by
aggregating
patterns
of
repetitions
of
that
structure,
as
well
as
variations
on
these
patterns.
The
mechanism
of
learning
is
based
on
priming:
when
multiple
repetitions
establish
and
reinforce
an
emotional
reflex
until
the
numerous
cultural
cues
adjust
the
emotional
reaction,
finalize,
and
fix
a
particular
emotional
state.
Once
this
emotional
response
is
reinforced
a
number
of
times,
this
state
becomes
conditioned
by
the
given
musical
structure.
It
can
still
be
modified
or
reconditioned
by
variations
in
the
conditioning
pattern,
if
these
variations
are
structurally
stable
enough
and
sufficiently
repeated.
According
to
Pribram,
this
kind
of
emotional
meaning
is
processed
in
parallel,
quite
independently
from
the
aesthetic
meaning,
which
is
related
mostly
to
recognition
of
variations
between
the
familiar
musical
structures.
Pribram
believes
that
the
reason
for
the
controversy
between
musicologists
as
to
what
constitutes
meaning
in
music
were
largely
because
the
scholars
looked
for
sources
of
meaning
in
things
that
relate
to
music
structure
by
reference,
via
cultural
convention,
whereas
in
truth
musical
structures
evoke
feeling
rather
than
referencing
them.
An
illustration
of
purely
emotional
meaning,
at
the
absence
of
aesthetic
evaluation
of
music,
would
be
what
is
called
in
psychomusicology
“evaluative
conditioning”
–
the
induction
of
emotion
by
purely
mechanical
pairing
of
a
specific
piece
of
music
with
a
specific
positive
or
negative
stimulus.
Over
time,
after
repeated
exposure,
this
pairing
will
become
fixed
into
a
“conditioned
stimulus,”
so
that
the
mere
entrance
of
familiar
pattern
of
sounds
will
evoke
the
conditioned
emotional
state
even
if
the
person
is
not
consciously
aware
of
music,
not
paying
attention
to
it,
or
is
busy
with
something
else.
Obviously,
in
such
a
case,
a
person
would
not
be
capable
of
executing
aesthetic
evaluation,
yet
he
will
stay
receptive
to
the
denotative
emotional
meaning
of
music
(Juslin
2011).
Pribram’s
insight
resolves
the
question
of
the
objectivity
of
music.
If
musical
denotation
is
made
by
means
of
direct
conditioning,
and
not
the
conscious
effort
of
memorizing
a
reference,
then
musical
denotation
is
undoubtedly
an
objective
process.
Just
as
the
smell
of
food
induces
the
emotion
of
hunger
–
musical
structures
elicit
specific
emotional
responses
in
a
listener.
The
objectivity
of
musical
emotion
is
further
supported
by
progress
in
the
automatic
recognition
of
perceived
emotion
in
music.
A
number
of
engineers,
starting
from
2000,
5
Of
course,
it
should
be
remembered
that
denotation
is
a
voluntary
act,
and
as
such,
learning
a
denotation
requires
an
individual’s
willingness
to
admit
the
convention
of
ascribing
a
particular
emotional
condition
to
a
particular
structural
element
of
music.
If
for
some
reason
an
individual
has
aversion
to
such
an
emotional
condition
or
musical
element,
the
denotation
mechanism
cannot
be
established.
This
is
in
contrast
to
the
musical
connotation,
which
seems
to
operate
semi-‐automatically.
20
have
been
working
to
build
applications
capable
of
organizing
musical
databases
according
to
the
emotional
content
of
music.
Automated
emotion
recognition
obviously
would
not
be
possible
unless
there
is
a
definable
set
of
emotional
markers
objectively
present
in
music.
Most
of
such
software
is
based
on
the
extraction
of
musical
attributes
from
a
live
stream
of
music
and
formulating
an
analytic
algorithm
for
the
computer
to
follow,
in
the
process
of
which
the
program
would
fine-‐tune
the
correlation
between
a
detected
structure
and
a
particular
emotional
label
provided
by
the
human
annotators
–
quite
similar
to
optical
character
recognition
(OCR)
(Yang
and
Chen
2011,
15–33).
Recently,
a
number
of
researchers
raised
an
interesting
question:
are
musical
emotions
really
felt?
Or,
do
listeners
simply
recognize
what
emotion
is
implied
by
music,
but
block
themselves
from
its
direct
experience?
Many
earlier
studies
did
not
account
in
their
experimental
setting
for
the
possibility
that
verbal
self-‐reports
of
musical
emotions
by
the
subjects
of
a
study
may
be
unreliable
–
because
listeners
confuse
the
emotions
expressed
in
the
music
with
their
own
emotions
in
their
answers
to
the
questionnaires.
Yet
another
problem
is
the
way
in
which
performers
handle
their
emotional
reactions
during
practicing
–
especially,
when
they
drill
“cold”
technical
use
of
specific
expressive
structures
(Gabrielsson
2002).
Of
course,
a
number
of
studies
have
avoided
this
pitfall
by
cross-‐referring
the
verbal
reports
with
the
data
of
physiological
responses
to
music.
Several
experiments
have
shown
that
music
listening
can
give
rise
to
physiological
responses
very
similar
to
those
shown
by
“natural”
emotional
stimuli,
such
as
physical
events.
The
reports
on
psychophysiological
responses
to
music
include
changes
in
heart
rate,
biochemical
responses,
skin
temperature,
skin
conductance,
respiration,
blood
pressure,
muscular
tension,
blood
volume,
gastric
mobility,
blood
oxygen,
and
pupillary
and
startle
reflexes
(Juslin
1993).
Different
pieces
of
music
can
produce
different
patterns
of
physiological
responses,
such
that
it
is
possible
to
discriminate
between
emotions
based
on
the
unique
configuration
of
psychophysiological
measurements
(Krumhansl
1997).
More
recent
studies
of
brain
activation
provide
additional
support
for
the
“reality”
of
emotional
reactions
to
music.
Listeners’
responses
to
music
activate
brain
regions
that
are
known
from
previous
studies
to
be
implicated
in
emotional
responses.
Functional
neuroimaging
and
lesion
studies
show
that
music-‐evoked
emotions
can
modulate
activity
in
virtually
all
limbic
and
paralimbic
brain
structures.
These
structures
are
crucially
involved
in
the
initiation,
generation,
detection,
maintenance,
regulation,
and
termination
of
emotions
(Koelsch
2010).
Further
evidence
that
listeners
actually
“feel”
musical
emotions
comes
from
research
on
behavior
influenced
by
exposure
to
music.
Observation
of
changes
of
facial
expressions
corroborated
with
electromyographic
(EMG)
measures
of
facial
muscles
and
measurement
of
heart
rate
fluctuations
throughout
listening
to
music
that
expresses
different
emotions,
leave
no
doubt
that
listeners
actually
feel
what
they
report.
The
negative
reaction
to
negative
emotions,
and
positive
reaction
to
positive
emotions
becomes
even
more
23
pronounced
with
longer
exposure
to
music
(Witvliet
and
Vrana
2007).
Lundqvist
et
al.
added
extra
measurements
of
skin
conductance
and
finger
temperature,
and
had
music
composed
specifically
for
the
study
to
avoid
possible
familiarity
of
the
subjects
with
the
music
that
could
have
affected
their
emotional
reactions.
The
results
confirmed
the
other
studies,
leading
to
the
conclusion
that
“emotions
induced
by
music
are
genuine
emotions
-‐
the
same
as
people
experience
in
real
life”
(Lundqvist
et
al.
2008).
More
specific
psychological
studies
based
on
questionnaires
addressed
the
issue
of
the
authenticity
of
musically
induced
emotions
by
directly
asking
the
subjects
what
they
felt
while
listening,
and
what
emotion
they
recognized
in
the
music.
For
the
most
part,
the
correspondence
between
the
two
was
high.
There
was,
however,
a
listener
bias:
positive
musical
emotions
were
reported
as
felt
more
often
than
recognized,
while
the
opposite
held
for
negative
emotions
(i.e.
fear).
This
may
reflect
most
listeners’
aesthetic
bias
–
they
identify
fear
as
a
negative
emotion
within
the
context
of
music,
but
regard
the
experience
of
“fear”
as
“positive”
in
relation
to
their
listening
experience
(Kallinen
and
Ravaja
2006).
This
is
not
unlike
people’s
enjoyment
of
horror
films
and
roller
coasters
which
elicit
strong
negative
emotions,
but
are
reported
as
“enjoyable”
activities.
Similar
results
were
obtained
when
the
subjects
were
asked
to
imagine
a
piece
of
music
and
report
the
emotional
content
of
the
imagined
sounds
(P.
Evans
and
Schubert
2008).
A
dedicated
interview-‐based
study,
designed
to
define
the
relationship
between
felt
and
performed
emotions
in
performing
musicians
while
preparing
for
their
performances,
revealed
that
musicians
use
different
strategies
to
manipulate
their
emotions.
In
the
first
stages
of
work,
they
distinguish
between
their
own
practice-‐related
emotions
(i.e.
frustration)
and
emotions
related
to
music
(i.e.
delight).
As
the
study
progresses,
they
cut
down
on
the
practice-‐related
emotions
and
focus
more
on
the
musical
emotions.
When
the
performance
plan
is
created,
the
“feeling”
of
musical
emotions
transforms
into
“knowing”
them.
As
the
performer
polishes
the
musical
expressions
in
preparation
for
a
concert,
the
intensity
of
felt
emotions
wanes.
Finally,
when
a
performer
comes
close
to
the
date
of
a
public
performance,
and
the
technicality
of
the
piece
is
mastered,
he
emphasizes
feeling
the
musical
emotions
once
more
(Van
Zijl
and
Sloboda
2011).
In
addition
to
the
emotions
directly
induced
by
music,
another
important
emotional
response
to
consider
is
the
listener’s
aesthetic
evaluation
of
the
music
they
hear.
A
number
of
composers,
critics,
and
musicologists
of
the
20th
century
contend
that
“beauty”
in
music
is
historically
limited,
and
no
longer
relevant
for
modern
art.
This
view
is
inconsistent
with
psychological
investigations
of
aesthetic
perception
of
music
by
performers,
for
instance,
which
confirm
that
musicians
evaluate
music
in
terms
of
“beauty”.
This
indicates
that
beauty
is
central
to
the
concept
of
aesthetic
response
to
music.
The
feature
of
being
“beautiful”
rates
the
highest
as
the
most
positive
of
the
adjectives
used
to
refer
to
the
value
of
music.
Not
surprisingly,
experiencing
music
as
“beautiful”
tends
to
enhance
the
listener’s
responses
to
the
expression
of
positive
musical
emotions.
However,
“beauty”
can
be
a
bit
more
nuanced
than
it
would
first
seem
-‐
“sad”
and
“beautiful”
can
co-‐occur
in
an
individual’s
estimation
of
a
musical
expression,
whereas
“ugly”
and
“beautiful”,
were
not
ever
reported
together
by
study
participants
(Van
Zijl
and
Sloboda
2011).
So,
there
are
three
distinct
emotional
phases
that
make-‐up
the
music
listening
experience:
24
(1) the
musical
emotion,
i.e.
the
denotative
expressive
idiom(s)
conveyed
by
the
music,
(2) a
listener’s
physiological
emotional
response
to
the
musical
emotion
(contingent
upon
the
listener
having
sufficient
emotional
priming
to
experience
it),
and
(3) the
emotional
response
the
listener
experiences
as
part
of
his
aesthetic
evaluation
of
the
musical
work.
For
example,
let’s
consider
a
piece
of
music
which
contains
a
period
(section
of
music)
with
a
very
melancholy
character
(1)
(eg.
“Adagio
for
Strings”
by
Barber).
The
listener
recognizes
this
music
as
“sad”,
and
emotes
to
it
by
experiencing
the
emotion
of
“sadness”
(2).
However,
the
listener’s
aesthetic
judgment
(3)
of
the
work
is
“positive”
–
so
he
considers
that
he
has
enjoyed
the
work,
despite
his
emotional
state
being
negative
throughout
the
listening.
The
lack
of
understanding
of
the
distinction
between
the
second
and
third
states
described
above
is
the
source
of
much
confusion
and
controversy
amongst
researchers
in
the
field
of
music
and
emotion.
It
seems
that
Karl
Pribram
was
generally
right
in
claiming
that
emotional
meaning
should
be
distinguished
from
aesthetic
meaning,
and
that
both
of
them
can
oppose
each
other.
Patrick
Juslin
also
finds
it
common
for
emotion
to
form
peculiar
relations
with
the
aesthetic
response.
Although
in
many
cases,
aesthetic
evaluation
follows
the
emotional
response,
a
piece
of
music
is
perfectly
capable
of
evoking
an
emotion
in
a
listener
without
triggering
any
aesthetic
evaluation.
This
is
especially
likely
in
relation
to
a
piece
of
popular
music,
which
usually
is
not
perceived
as
an
“art
object.”
On
another
hand,
it
is
possible
to
prefer
a
piece
of
music
over
some
other
piece,
without
the
music
arousing
any
emotion
–
just
by
the
virtue
of
liking
the
lyrics
of
that
song
or
the
sound
of
a
particular
musical
instrument
(Juslin
2011).
Altogether,
the
intricacies
of
distinguishing
emotional
and
aesthetic
meanings
in
music
are
no
more
complex
than
in
speech.
Poetic
speech
presents
exactly
the
same
problem:
the
literal
meaning
of
words
in
a
poem
cannot
be
taken
literally
without
consulting
the
aesthetic
aspect
of
verbal
expression.
And
the
greater
the
aesthetic
value
in
poetic
expression
is,
the
less
clear
the
exact
meaning
of
the
words
is.
The
opposite
is
also
true:
the
more
clear
the
verbal
expression,
usually
the
less
value
it
bears
from
an
artistic
standpoint.
Emotional
communication
through
music
pretty
much
follows
the
same
route
as
conceptual
communication
through
verbal
speech.
Both
are
designed
to
facilitate
coordination
between
an
individual
and
a
group.
Where
verbal
language
focuses
on
conveying
factual
data,
music‘s
purpose
is
to
communicate
emotions.
This
narrow
specialization
determines
the
entire
infrastructure
in
both,
music
and
verbal
language.
• The
context
of
verbal
language
serves
to
establish
the
framework
of
reference
for
concrete
objects,
their
qualities,
actions,
and
events.
• The
context
of
music
reflects
on
emotional
states,
utilizing
representations
of
qualities
and
actions
to
the
extent
necessary
to
establish
a
correlation
between
conceptual
and
emotional
modes
of
thinking.
The
way
in
which
verbal
language
encourages
conceiving
ideas
to
reflect
on
reality,
music
promotes
the
generation
of
emotions.
Both
verbal
and
musical
modes
form
two
sides
of
the
same
communication
coin.
The
development
of
both
the
conceptual
and
emotional
25
faculties
is
vital
for
the
survival
of
an
individual.
What
makes
music
special
is
its
greater
accessibility:
due
to
its
reliance
on
genetically
embedded
sound
gestures
rather
than
referential
conventions
of
meaning,
it
reaches
more
people
across
wider
cultural
borders.
So
we
can
safely
conclude,
then,
that
the
common
figure
of
speech
is
literally
true:
music
is
the
universal
language
of
emotion.
REFERENCES:
Bachorik,
Justin
Pierre,
Marc
Bangert,
Psyche
Loui,
Kevin
Larke,
Jeff
Berger,
Robert
Rowe,
and
Gottfried
Schlaug.
2009.
“Emotion
in
Motion:
Investigating
the
Time-‐Course
of
Emotional
Judgments
of
Musical
Stimuli.”
Music
Perception:
An
Interdisciplinary
Journal
26
(4):
355–64.
doi:10.1525/mp.2009.26.4.355.
Bowling,
Daniel
Liu,
Janani
Sundararajan,
Shui’er
Han,
and
Dale
Purves.
2012.
“Expression
of
Emotion
in
Eastern
and
Western
Music
Mirrors
Vocalization.”
Edited
by
Olaf
Sporns.
PLoS
ONE
7
(3):
e31942.
doi:10.1371/journal.pone.0031942.
Clarke,
Eric
F.
2001.
“Meaning
and
the
Specification
of
Motion
in
Music.”
Musicae
Scientiae
5
(2):
213–34.
doi:10.1177/102986490100500205.
Damasio,
Antonio.
2012.
Self
Comes
to
Mind:
Constructing
the
Conscious
Brain.
New
York:
Vintage
Books.
https://books.google.com/books?id=PBH40-‐TgBCgC.
de
Waal,
Frans
B.
M.
2011.
“What
Is
an
Animal
Emotion?”
Annals
of
the
New
York
Academy
of
Sciences
1224
(April):
191–206.
doi:10.1111/j.1749-‐6632.2010.05912.x.
Dolscheid,
S.,
S.
Hunnius,
D.
Casasanto,
and
A.
Majid.
2014.
“Prelinguistic
Infants
Are
Sensitive
to
Space-‐Pitch
Associations
Found
Across
Cultures.”
Psychological
Science
25
(6):
1256–61.
doi:10.1177/0956797614528521.
Draghi-‐Lorenz,
Riccardo,
Vasudevi
Reddy,
and
Alan
Costall.
2001.
“Rethinking
the
Development
of
‘Nonbasic’
Emotions:
A
Critical
Review
of
Existing
Theories.”
Developmental
Review
21
(3):
263–304.
doi:10.1006/drev.2000.0524.
Egermann,
H.,
M.
E.
Sutherland,
O.
Grewe,
F.
Nagel,
R.
Kopiez,
and
E.
Altenmuller.
2011.
“Does
Music
Listening
in
a
Social
Context
Alter
Experience?
A
Physiological
and
Psychological
Perspective
on
Emotion.”
Musicae
Scientiae
15
(3):
307–23.
doi:10.1177/1029864911399497.
Egermann,
Hauke,
and
Stephen
McAdams.
2013.
“Empathy
and
Emotional
Contagion
as
a
Link
Between
Recognized
and
Felt
Emotions
in
Music
Listening.”
Music
Perception:
An
Interdisciplinary
Journal
31
(2):
139–56.
doi:10.1525/mp.2013.31.2.139.
Eitan,
Zohar,
and
Roni
Y.
Granot.
2006.
“How
Music
Moves.”
Music
Perception
23
(3):
221–48.
doi:10.1525/mp.2006.23.3.221.
Ekman,
Paul.
2003.
Emotions
Revealed:
Recognizing
Faces
and
Feelings
to
Improve
Communication
and
Emotional
Life.
London:
Macmillan.
https://books.google.com/books?id=JJ9xHXo7hPgC.
Evans,
K.
K.,
and
A.
Treisman.
2011.
“Natural
Cross-‐Modal
Mappings
between
Visual
and
Auditory
Features.”
Journal
of
Vision
10
(1):
6–6.
doi:10.1167/10.1.6.
Evans,
P.,
and
E.
Schubert.
2008.
“Relationships
between
Expressed
and
Felt
Emotions
in
Music.”
Musicae
Scientiae
12
(1):
75–99.
doi:10.1177/102986490801200105.
Feldman,
Jacob,
David
Epstein,
and
Whitman
Richards.
1992.
“Force
Dynamics
of
Tempo
Change
in
Music.”
Music
Perception:
An
Interdisciplinary
Journal
10
(2):
185–203.
doi:10.2307/40285606.
Fritz,
Thomas,
Daniela
Sammler,
and
Stefan
Koelsch.
2006.
“How
Far
Is
Music
Universal?
An
Intercultural
Comparison.”
In
9th
International
Conference
on
Music
Perception
&
Cognition,
Bologna,
Italy,
edited
by
M.
Baroni,
A.
R.
Addessi,
R.
Caterina,
and
M.
Costa,
88.
Bologna,
Italy:
26
Bolonia
University
Press.
Gabrielsson,
Alf.
2002.
“Emotion
Perceived
and
Emotion
Felt:
Same
or
Different?”
Musicae
Scientiae
5
(1
suppl):
123–47.
doi:10.1177/10298649020050S105.
Gabrielsson,
Alf,
and
Patrik
N.
Juslin.
2003.
“Emotional
Expression
in
Music.”
In
Handbook
of
Affective
Sciences,
edited
by
R.
J.
Davidson,
K.
R.
Scherer,
and
H.
H.
Goldsmith,
503–34.
Series
in
Affective
Science.
New
York,
NY,
US:
Oxford
University
Press.
Gabrielsson,
Alf,
and
Siv
Lindström
Wik.
2003.
“Strong
Experiences
Related
to
Music:
Adescriptive
System.”
Musicae
Scientiae
7
(2):
157–217.
doi:10.1177/102986490300700201.
Hubbard,
Timothy
L.
1996.
“Synesthesia-‐like
Mappings
of
Lightness,
Pitch,
and
Melodic
Interval.”
The
American
Journal
of
Psychology
109
(2):
219–38.
Humrichouse,
John,
Michael
Chmielewski,
Elizabeth
A.
McDade-‐Montez,
and
David
Watson.
2007.
“Affect
Assessment
Through
Self-‐Report
Methods.”
In
Emotion
and
Psychopathology:
Bridging
Affective
and
Clinical
Science,
edited
by
J.
Rottenberg
and
S.
L.
Johnson,
13–34.
Washington,
DC,
US:
American
Psychological
Association.
Huron,
David,
and
Jonathon
Berec.
2009.
“Characterizing
Idiomatic
Organization
in
Music:
A
Theory
and
Case
Study
of
Musical
Affordances.”
Empirical
Musicology
Review
4
(3):
103–22.
http://hdl.handle.net/1811/44531.
Juslin,
Patrik
N.
1993.
“Psychophysiological
Measures.”
In
Handbook
of
Music
and
Emotion:
Theory,
Research,
Applications,
edited
by
Patrik
N.
Juslin
and
John
A.
Sloboda,
279–311.
Oxford,
UK:
Oxford
University
Press.
———.
2005.
“From
Mimesis
to
Catharsis.”
In
Musical
Communication,
edited
by
Dorothy
Miell,
Raymond
MacDonald,
and
David
J.
Hargreaves,
85–116.
Oxford,
UK:
Oxford
University
Press.
———.
2011.
“Music
and
Emotion:
Seven
Questions,
Seven
Answers.”
In
Music
and
the
Mind,
edited
by
Irène
Deliège
and
Jane
Davidson,
113–38.
Oxford,
UK:
Oxford
University
Press.
http://www.oxfordscholarship.com/view/10.1093/acprof:osobl/9780199581566.001.0001/
acprof-‐9780199581566-‐chapter-‐7.
Juslin,
Patrik
N.,
and
Petri
Laukka.
2003.
“Communication
of
Emotions
in
Vocal
Expression
and
Music
Performance:
Different
Channels,
Same
Code?”
Psychological
Bulletin
129
(5):
770–814.
doi:10.1037/0033-‐2909.129.5.770.
Kagan,
Jerome.
2007.
What
Is
Emotion?:
History,
Measures,
and
Meanings.
New
Haven,
CT:
Yale
University
Press.
Kallinen,
K.,
and
N.
Ravaja.
2006.
“Emotion
Perceived
and
Emotion
Felt:
Same
and
Different.”
Musicae
Scientiae
10
(2):
191–213.
doi:10.1177/102986490601000203.
Kim,
Jung-‐Kyong,
and
Robert
J.
Zatorre.
2010.
“Can
You
Hear
Shapes
You
Touch?”
Experimental
Brain
Research
202
(4):
747–54.
doi:10.1007/s00221-‐010-‐2178-‐6.
Koelsch,
Stefan.
2010.
“Towards
a
Neural
Basis
of
Music-‐Evoked
Emotions.”
Trends
in
Cognitive
Sciences
14
(3):
131–37.
doi:10.1016/j.tics.2010.01.002.
Krumhansl,
C.
L.
1997.
“An
Exploratory
Study
of
Musical
Emotions
and
Psychophysiology.”
Canadian
Journal
of
Experimental
Psychology
=
Revue
Canadienne
De
Psychologie
Expérimentale
51
(4):
336–53.
http://www.ncbi.nlm.nih.gov/pubmed/9606949.
Kwoun,
S.-‐J.
2009.
“An
Examination
of
Cue
Redundancy
Theory
in
Cross-‐Cultural
Decoding
of
Emotions
in
Music.”
Journal
of
Music
Therapy
46
(3):
217–37.
doi:10.1093/jmt/46.3.217.
Lehmann,
Andreas
C.,
John
A.
Sloboda,
and
Robert
H.
Woody.
2007.
Psychology
for
Musicians :
Understanding
and
Acquiring
the
Skills.
Oxford,
UK:
Oxford
University
Press,
USA.
Lipscomb,
Scott
D;,
and
Donald
Hodges.
1996.
“Hearing
and
Music
Perception.”
In
Handbook
of
Music
Psychology,
edited
by
Donald
Hodges,
2nd
ed.,
83–132.
San
Antonio,
TX:
IMR
Press.
27
Logeswaran,
Nidhya,
and
Joydeep
Bhattacharya.
2009.
“Crossmodal
Transfer
of
Emotion
by
Music.”
Neuroscience
Letters
455
(2):
129–33.
doi:10.1016/j.neulet.2009.03.044.
Lundqvist,
L.-‐O.,
F.
Carlsson,
P.
Hilmersson,
and
P.
N.
Juslin.
2008.
“Emotional
Responses
to
Music:
Experience,
Expression,
and
Physiology.”
Psychology
of
Music
37
(1):
61–90.
doi:10.1177/0305735607086048.
Marks,
Lawrence
E.
1978.
The
Unity
of
the
Senses:
Interrelations
among
the
Modalities.
New
York:
Academic
Press.
———.
2014.
“Synesthesia.”
In
Varieties
of
Anomalous
Experience:
Examining
the
Scientific
Evidence,
edited
by
Etzel
Cardena,
Steven
J.
Lynn,
and
Stanley
Krippner,
121–49.
Washington
DC:
American
Psychological
Association.
https://books.google.com/books?id=BRbDnQEACAAJ.
Matsumoto,
D.,
and
H.
S.
Hwang.
2012.
“Culture
and
Emotion:
The
Integration
of
Biological
and
Cultural
Contributions.”
Journal
of
Cross-‐Cultural
Psychology
43
(1):
91–118.
doi:10.1177/0022022111420147.
Matsumoto,
D.,
H.
S.
Hwang,
and
H.
Yamada.
2012.
“Cultural
Differences
in
the
Relative
Contributions
of
Face
and
Context
to
Judgments
of
Emotions.”
Journal
of
Cross-‐Cultural
Psychology
43
(2):
198–218.
doi:10.1177/0022022110387426.
Milstein,
Yakov.
1971.
Liszt.
2nd
ed.
Vol.
2.
Moscow:
Muzyka.
Mohn,
C.,
H.
Argstatter,
and
F.-‐W.
Wilker.
2011.
“Perception
of
Six
Basic
Emotions
in
Music.”
Psychology
of
Music
39
(4):
503–17.
doi:10.1177/0305735610378183.
Nykänen,
Arne.
2004.
Methods
for
Specification
of
Sound
Quality
Applied
to
Saxophone
Sound.
Luleå,
Sweden:
Luleå
University
of
Technology.
http://epubl.ltu.se/1402-‐1757/2004/70/LTU-‐LIC-‐
0470-‐SE.pdf.
Palmer,
S.
E.,
K.
B.
Schloss,
Z.
Xu,
and
L.
R.
Prado-‐Leon.
2013.
“Music-‐Color
Associations
Are
Mediated
by
Emotion.”
Proceedings
of
the
National
Academy
of
Sciences
110
(22):
8836–41.
doi:10.1073/pnas.1212562110.
Panksepp,
Jaak.
1998.
Affective
Neuroscience:
The
Foundations
of
Human
and
Animal
Emotions.
Oxford,
UK:
Oxford
University
Press.
Parise,
Cesare
Valerio,
and
Charles
Spence.
2009.
“‘When
Birds
of
a
Feather
Flock
Together’:
Synesthetic
Correspondences
Modulate
Audiovisual
Integration
in
Non-‐Synesthetes.”
Edited
by
David
C.
Burr.
PLoS
ONE
4
(5):
e5664.
doi:10.1371/journal.pone.0005664.
Peretz,
Isabelle.
2013.
“Towards
a
Neurobiology
of
Musical
Emotions.”
In
Handbook
of
Music
and
Emotion:
Theory,
Research,
Applications,
edited
by
Patrik
N.
Juslin
and
John
A.
Sloboda,
99–126.
Oxford,
UK:
OUP
Oxford.
doi:10.1093/acprof:oso/9780199230143.003.0005.
Persson,
R.S.
2010.
“Adding
Emotion
to
the
Gifted
Musical
Mind:
Towards
a
Model
of
Musical
Thinking.”
The
International
Journal
of
Creativity
and
Problem
Solving
20
(2).
The
International
Journal
of
Creativity
and
Problem
Solving:
85–107.
Pribram,
Karl
H.
1982.
“Brain
Mechanism
in
Music:
Prolegomena
for
a
Theory
of
the
Meaning
of
Meaning.”
In
Music,
Mind,
and
Brain:
The
Neuropsychology
of
Music,
edited
by
Manfred
Clynes,
21–35.
New
York:
Plenum
Press.
https://books.google.com/books?id=XqPVBwAAQBAJ.
Reid,
Stefan.
2002.
“Preparing
for
Performance.”
In
Musical
Performance:
A
Guide
to
Understanding,
edited
by
John
Rink,
102–12.
Cambridge,
UK:
Cambridge
University
Press.
Russo,
Frank
A.,
Paolo
Ammirante,
and
Deborah
I.
Fels.
2012.
“Vibrotactile
Discrimination
of
Musical
Timbre.”
Journal
of
Experimental
Psychology.
Human
Perception
and
Performance
38
(4):
822–26.
doi:10.1037/a0029046.
Saeed,
John
I.
2015.
Semantics.
Oxford,
UK:
John
Wiley
&
Sons.
Sievers,
Beau,
Larry
Polansky,
Michael
Casey,
and
Thalia
Wheatley.
2013.
“Music
and
Movement
Share
a
Dynamic
Structure
That
Supports
Universal
Expressions
of
Emotion.”
Proceedings
of
28
the
National
Academy
of
Sciences
of
the
United
States
of
America
110
(1):
70–75.
doi:10.1073/pnas.1209023110.
Simner,
Julia.
2012.
“Defining
Synaesthesia:
A
Response
to
Two
Excellent
Commentaries:
Defining
Synaesthesia.”
British
Journal
of
Psychology
103
(1):
24–27.
doi:10.1111/j.2044-‐
8295.2011.02059.x.
Sloboda,
J.
A.
1991.
“Music
Structure
and
Emotional
Response:
Some
Empirical
Findings.”
Psychology
of
Music
19
(2):
110–20.
doi:10.1177/0305735691192002.
Stravinsky,
Igor.
1936.
Autobiography.
Translated
by
Walter
Nouvel.
Chroniques
de
Ma
Vie.
New
York:
Simon
and
Schuster.
———.
2005.
Chroniques
de
Ma
Vie
[Хроника
моей
музыкальной
жизни].
Edited
by
I.
Vershinina.
Translated
by
V.M.
Bogdanov-‐Berezovskii.
Moscow:
Kompozitor
[Композитор].
Thompson,
W.
F.,
and
L.-‐L.
Balkwill.
2010.
“Cross-‐Cultural
Similarities
and
Differences.”
In
Handbook
of
Music
and
Emotion:
Theory,
Research,
Applications
.,
755–88.
Oxford,
UK:
Oxford
University
Press.
doi:10.1037/e512802011-‐003.
Tsai,
Chen-‐Gia,
Li-‐Ching
Wang,
Shwu-‐Fen
Wang,
Yio-‐Wha
Shau,
Tzu-‐Yu
Hsiao,
and
Wolfgang
Auhagen.
2010.
“Aggressiveness
of
the
Growl-‐Like
Timbre:
Acoustic
Characteristics,
Musical
Implications,
and
Biomechanical
Mechanisms.”
Music
Perception
27
(3):
209–22.
doi:10.1525/mp.2010.27.3.209.
Tsang,
Tawny,
and
Karen
B.
Schloss.
2012.
“Associations
between
Color
and
Music
Are
Mediated
by
Emotion
and
Influenced
by
Tempo.”
The
Yale
Review
of
Undergraduate
Research
in
Psychology
82:
82–93.
Van
Zijl,
A.
G.
W.,
and
J.
Sloboda.
2011.
“Performers’
Experienced
Emotions
in
the
Construction
of
Expressive
Musical
Performance:
An
Exploratory
Investigation.”
Psychology
of
Music
39
(2):
196–219.
doi:10.1177/0305735610373563.
Watt,
Roger
J.,
and
Roisin
L.
Ash.
1998.
“A
Psychological
Investigation
of
Meaning
in
Music.”
Musicae
Scientiae
2
(1):
33–53.
doi:10.1177/102986499800200103.
Weinstein,
Bryan,
and
Mark
C.
Gridley.
2010.
“Visual
Perception
of
Music.”
Psychology
Journal
7
(3):
80–87.
Witherington,
David
C.,
Joseph
J.
Campos,
Jennifer
A.
Harriger,
Cheryl
Bryan,
and
Tessa
E.
Margett.
2010.
“Emotion
and
Its
Development
in
Infancy.”
In
The
Wiley-‐Blackwell
Handbook
of
Infant
Development,
edited
by
J.
Gavin
Bremner
and
Theodore
D.
Wachs,
568–91.
Oxford,
UK:
Wiley-‐
Blackwell.
doi:10.1002/9781444327564.ch19.
Witvliet,
Charlotte
V.
O.,
and
Scott
R.
Vrana.
2007.
“Play
It
Again
Sam:
Repeated
Exposure
to
Emotionally
Evocative
Music
Polarises
Liking
and
Smiling
Responses,
and
Influences
Other
Affective
Reports,
Facial
EMG,
and
Heart
Rate.”
Cognition
&
Emotion
21
(1):
3–25.
doi:10.1080/02699930601000672.
Yampolskii,
Izrail
Y.
1961.
Niccolo
Paganini:
Life
and
Career
[Никколо
Паганини.
Жизнь
и
творчество].
Moscow:
Gos
Muz
Izdat
[Гос.
музыкальное
изд-‐во].
Yang,
Yi-‐Hsuan,
and
Homer
H.
Chen.
2011.
Music
Emotion
Recognition.
Boca
Raton,
FL:
CRC
Press.