You are on page 1of 4

2020 13th International Symposium on Computational Intelligence and Design (ISCID)

Exploiting the emotional preference of music for music recommendation


in daily activities
2020 13th International Symposium on Computational Intelligence and Design (ISCID) | 978-1-7281-8446-3/20/$31.00 ©2020 IEEE | DOI: 10.1109/ISCID51228.2020.00085

Hui Zhang1 , Kejun Zhang1, ∗ , Nick Bryan-Kinns2


1 Laboratory
of CAD, CG, Zhejiang University, China
2 School of Electronic Engineering and Computer Science, Queen Mary University of London, the United Kingdom
{huiz, zhangkejun}@zju.edu.cn, n.bryan-kinns@qmul.ac.uk

Abstract—The emotional use of music plays an impor- systems to obtain accurate user emotional information [3].
tant role in influencing people’s music-searching behaviour. Through reviewing the work of these predecessors, it can
Many users listen to music for multiple mood regulation be seen that currently the common processes of research on
purposes. Therefore, emotional preference should be taken into
consideration especially when building context-aware music emotion-aware music recommendation is to first reason user
recommendation systems which aim to satisfy users’ short-term emotions and then recommend music with similar emotional
music needs. In addition, cultural differences will affect the way states based on the obtained emotional cues. However, in
people experience and value emotion from music. Based on the case of self-chosen music listening, people’s emotional
these aspects, we investigate the emotional preference of music needs for music are not necessarily in line with their real-
in daily activities through a cross-cultural survey in China and
the UK, and build emotional maps between activity and music time emotional state. For example, when someone gets
for the two countries. Furthermore, we introduce EmoMusic, up in the morning and still feel sleepy, he may prefer
a novel emotion-based music recommendation service for daily to have some energetic songs playing instead of drowsy
activities which supports visualisation and control of music songs. Mood management and regulation is one of the most
emotion through an interactive interface. A user study is pre- common motivations and functions of self-chosen music
sented to evaluate the application. Results reveal good usability
of the service and show that visualization and interaction listening behaviour [4], and the emotional uses of music
in a music emotion space can improve acceptance of music vary across different daily activities [5]. Considering this
recommendations. Our work helps to understand the emotional aspect, the authors suggest that the emotion-aware music
preference of music for various activities and how to exploit recommendation systems should not only consider the real-
emotional cues of music in a music recommendation service, time emotional state of users, but should recommend songs
paving the way for more user-centered music recommendation
systems. that can prompt users to reach their target emotional state.
It leads to the first research question in this paper:
Keywords-user-centered music recommendation; emotional
preference; cross-cultural RQ1. Will the user-expected musical emotion match their
emotional state when users start listening to the song
I. I NTRODUCTION (initial emotion state), or do users prefer music which
As one of the key factors that affect music listening can facilitate achieving another emotional state (target
behaviors of human, emotion has played an important role emotion state)?
in studies on context-aware music recommendation systems. Furthermore, cultural differences may exist in emotional
Park et al. [1] use a fuzzy Bayesian network to deduce a uses of music. Previous studies have shown that emotion is
user’s mood from context information including weather, not only biologically determined, but is also affected by the
noise, time, gender as well as age, and then recommend social environment. The way people experience and value
music that matches the deduced user mood using manually the emotion varies significantly across different cultural
annotated emotion labels of each available song. This is backgrounds [6]–[8]. Therefore, it’s necessary to explore
probably the first research work which proposes the concept how cultural and national context may affect people’s feel-
of context-aware music recommendation. Chen-Ya’s work ings about the emotional uses and preference of music,
[2] on context and emotion aware system for personalized which leads to the following research question:
music recommendation also measures the similarity between RQ2. Will British people and Chinese people have different
a user’s emotion and the emotional character of a song preferences for musical emotions in each daily activity?
as the affective indicator. In addition, with the develop-
This paper sets out to exploit the emotional use of music
ment of wearable physiological sensors, some researchers
in various daily activities in order to optimize the experience
have tried integrating emotion recognition technique based
of user-centered, context-aware music recommendation. The
on physiological signals into the music recommendation
authors have conducted a cross-cultural study to explore peo-
∗ Kejun Zhang is the corresponding author. ple’s music preference in daily activities and implemented

2473-3547/20/$31.00 ©2020 IEEE 350


DOI 10.1109/ISCID51228.2020.00085

Authorized licensed use limited to: University of Prince Edward Island. Downloaded on May 26,2021 at 11:43:16 UTC from IEEE Xplore. Restrictions apply.
a web application built on the results. More specifically, the
major contributions of this paper are:
• Emotional preference of music in daily activities: We
explore emotional uses of music in daily activities both
in China and in the United Kingdom, which allows us
to build the mapping between music emotion and daily
activities as well as the cross-cultural comparison.
• Implementation and evaluation of EmoMusic: We im-
plement EmoMusic, a novel emotion-based music rec- Figure 1. The distribution of the scores of participants in the Gold-MSI
subscale of emotion factor (a) and the number of times each activity was
ommendation service for daily activities, which sup- chosen (b).
ports explanation and control of music emotion through
an interactive interface. In addition, we present a user Table I
study and discuss the results to demonstrate its perfor- R ESULTS OF TOST TESTS ON EXPECTED MUSIC EMOTION AND
TARGETING USER EMOTION
mance and usability.
China UK
II. O NLINE S URVEY
Valence Arousal Valence Arousal
In order to better understand emotional preferences of df 211 211 195 195
music listening in everyday situations and the influence of t(ΔL =−0.3) 3.526 3.403 7.353 5.293
cultural aspects, an online survey in Chinese and English p(ΔL =−0.3) < 0.001 < 0.001 < 0.001 < 0.001
t(ΔU =0.3) −6.309 −7.811 −5.401 −5.116
was designed and released from March to May 2020, in p(ΔU =0.3) < 0.001 < 0.001 < 0.001 < 0.001
China and the UK.
A. Participants
happy, 1 = calm and 5 = excited). There was an additional
A total of 89 participants took part in this survey, includ-
free response field for comments and advice. The average
ing 47 from China (29 females and 18 males, aging from
time to complete the survey for each participant is 17.70
18 to 37) and 42 from the UK (15 females and 27 males,
minute.
aging from 18 to 56). The minimum sample size (38 for two-
tailed paired t-test, dz = 0.5, α = 0.05, and power = 0.85) C. Results and Discussion
of participants in each country was calculated using prospec- Scores of participants for the Gold-MSI subscale of emo-
tive power analysis before the experiment to guarantee the tion factor were calculated and can be seen in Figure 1a.
statistical defensibility of further analysis. Participants in both of China and the UK achieved compar-
B. Procedure atively high average scores (0.779 and 0.817 respectively),
which means the major proportion of participants are used
One page of research brief and a consent form were first
to consciously listening to music to manage their mood
presented online to participants before they decided to attend
in everyday life. Figure 1b shows the distribution of the
the survey. After that, participants were asked to fill in some
number of times each activity was selected. It can be seen
basic demographic information (gender, age and nationality)
that the most of activities share similar popularity except
and six items from the Gold-MSI subscale of emotion
that partying, waiting and social interaction differ greatly
factor (importance of music for emotional functions) [9].
in ranking (> 3) between two countries. Though there is
Participants were next required to select five daily activities
some difference in the number of times choosing activities
when they will listen to music from 13 situations shown in
between the two country, there is no significant difference
Table II. For each activity, participants were asked to answer
in the importance and the frequency of listening to music
questions about the frequency (Q1) and importance (Q2) of
in these activities according to the answers of Q1 and Q2.
listening to music, as well as their emotional preference
It is interesting to find that participants in China and the
of music (Q3) and affective states before (Q4) /during
UK share similar levels of engaging in incidental musical
(Q5) this activity. In this paper, emotion is represented by
listening in everyday activities.
using the valence-arousal model (VA model) proposed by
One-way ANOVA analysis was performed on the user-
Russell [10], which is one of the most popular dimensional
expected musical emotion, initial emotion state and target
emotion conceptualizations and has been adopted by a large
emotion state in each daily activity. In both of countries,
proportion of Music Emotion Recognition (MER) research.
results revealed significant difference across valence of three
Q1 and Q2 were rated on a 5-point Likert scale (respectively,
kinds (fv = 65.663, pv < 0.01 in China, and fv =
1 = never and 5 = always, 1 = not at all and 5 = absolutely
55.873, pv < 0.01 in the UK) as well as arousal (fa =
essential). Valence and arousal in Q3, Q5 and Q6 were rated
31.796, pa < 0.01 in China, and fa = 9.265, pa < 0.01 in
on a 5-point Likert scale (respectively, 1 = sad and 5 =
the UK). In order to further analyze the pairwise difference

351

Authorized licensed use limited to: University of Prince Edward Island. Downloaded on May 26,2021 at 11:43:16 UTC from IEEE Xplore. Restrictions apply.
Table II
AVERAGED EMOTIONAL VALUES OF MUSIC FOR EACH ACTIVITY

Activity music valence music arousal emotional use


China UK China UK China UK
work/study 3.625 3.906 3.250* 2.531* +V+A +V
falling asleep 3.375 3.667 2.125 2.333 -A +V-A
waking up 4.083 4.000 4.167* 2.400* +V+A
commuting 3.931 4.000 3.759 3.688 +V+A +V+A
social interaction 3.500 3.833 3.500 3.667
housework 4.000 3.765 3.944 3.529 +V+A +V+A
physical activity 4.207 3.968 4.655 4.290 +V+A +V+A
shopping 4.500 3.200 3.500 2.400
having a meal 3.400 3.000 3.000 1.500 Figure 2. Screenshot of the EmoMusic interface. (a: left) pane for selecting
partying 4.571 4.474 4.286 4.211 +V daily activities, (b: centre) visualization of music emotion space, (c: right)
gaming 3.833 3.800 4.083 3.600 +V+A +V+A recommendation song list.
waiting 3.857 3.750 3.286 3.750 +V
relaxation 3.923 3.870 3.231* 2.565* +V +V

* means significant difference on this pair of emotional items of web application, EmoMusic, was developed and deployed
the two countries. online. The interaction of EmoMusic is designed based on
Note. Emotional use is presented only when there is significant simplicity and usability and its user interface consists of
difference between initial emotion and target emotion for this a navigation bar and a main operation area which can be
activity, e.g, +V means there is a significant increase from initial divided into three sections: activity selection (Figure 2a),
valence to target valence.
emotion visualization (Figure 2b) and recommendation panel
(Figure 2c). Due to the personalization of emotional needs
among the three sets of emotion value, we conducted t- and the dynamic nature of the listening context, adaptivity of
tests on them in pairs and found that mood change expected music recommendation service is very important [11]. Emo-
by participants was significant (tv = 12.388, pv < 0.01, Music implements the change of user emotional preference
ta = 8.352, pa =< 0.01 in China and tv = 10.234, pv < by enabling the movement of the target emotion star within
0.01, ta = 5.190, pa < 0.01 in the UK), while there the visualization panel, and the recommended songs will
was no significant difference between music emotion and be recalculated according to mood-based similarity. Music
target emotion. We further conducted an equivalence test collection used in EmoMusic are obtained from the PMEmo
on music emotion and target emotion using two-one-sided Dataset [12], which provides emotional annotations in both
t-tests (TOST). Results suggested that the expected music of valence and arousal dimension for 794 music excerpts.
emotion and targeting user emotion (both of valence and
A. User Study
arousal) were statistically equivalent (shown in Table I),
which means participants would like to listen to music of In order to evaluate the usability and performance of
emotion that matches the target emotion instead of their EmoMusic in recommending music for daily activities, we
initial emotional state. conducted a user study in the UK and invited 31 participants
The detailed values of expected music emotion for each (16 females and 15 males) from the age of 24 to 56.
activity are described in Table II, which can be utilized Participants were first asked to provide demographic related
to build emotional map between music emotion and daily information in a pre-study, then performed the main task
activities in China and the UK, as well as implement and finally gave qualitative feedback in a post study. During
emotion-based music recommendation system for everyday the main task, participant were asked to select a daily
life. In addition, a set of repeated t-tests was performed on activity they would like to do whilst listening to music.
initial emotion and target emotion to explore the emotional The emotional values of the selected activity were used
use of music listening in each activity. As shown in Table II, to generate a list of 5 recommended music tracks which
participants in China and the UK share similar expectation of were shown on the right side of the web page. Ratings were
mood management direction for the majority of daily activi- collected for 5 tracks in an initial recommendation list. After
ties though some difference exists, e.g., Chinese participants that, participants were allowed to interact freely with the
would like use music to increase both of valence and arousal application to generate recommendation lists as they wished.
when doing ‘work/study’ while UK participants only expect Once satisfied, they were required to again rate the full list
valence improvement. of tracks for the final list. In addition to rating individual
tracks, participants were also asked to provide overall ratings
III. E MO M USIC S ERVICE for the initial list and the final list.
To demonstrate how to utilize the emotional map between Distribution of scores rated by participants in this study
activity and music acquired from the previous survey to build is shown in Figure 3. As shown in Figure 3a, the average
emotion-based music recommendation service, a prototype rating of individual tracks was 3.323 for the initial playlist

352

Authorized licensed use limited to: University of Prince Edward Island. Downloaded on May 26,2021 at 11:43:16 UTC from IEEE Xplore. Restrictions apply.
the relation between personal characteristics and emotional
preference of music. (ii) incorporate emotional preference
with collaborative or content based recommendation engine
for better personalization.
ACKNOWLEDGMENT
This project is supported by the Key Project of Na-
tional Science Foundation of Zhejiang Province (No.
Figure 3. Scores rated by participants. (a) RI is rating of individual LZ19F020002), the National Natural Science Foundation
tracks in initial playlist, RF is rating of individual tracks in final playlist,
ORI is overall rating of initial playlist, and ORF is overall rating of final of China (No. 61672451), Provincial Key Research and
playlist. (b) VBU and VAU denote scores of valence before and after using Development Plan of Zhejiang Province (No. 2019C03137).
EmoMusic respectively, ABU and AAU denote scores of arousal before And the first author would like to thank the financial support
and after using EmoMusic respectively. (c) presents scores of level of
satisfaction and difficulty of using EmoMusic. The purple full lines inform of China Scholarship Council (CSC).
the median value and blue dotted lines inform the mean value.
R EFERENCES
[1] H.-S. Park, J.-O. Yoo, and S.-B. Cho, “A context-aware music
and 3.884 for the final playlist. Meanwhile, the mean value recommendation system using fuzzy bayesian networks with
of overall rating was 3.226 for the initial playlist and 3.839 utility theory,” in International conference on Fuzzy systems
and knowledge discovery. Springer, 2006, pp. 970–979.
for the final list, which indicated that participants were able [2] C.-Y. Wang, Y.-C. Wang, and S.-C. T. Chou, “A context and
to find better tracks through the interactive panel of music emotion aware system for personalized music recommenda-
emotion (t = 5.022, p < 0.01 for individual tracks and tion,” Journal of Internet Technology, vol. 19, no. 3, pp. 765–
t = 2.906, p < 0.01 for overall playlist). In order to detect 779, 2018.
the influence of EmoMusic on mood regulation, participants [3] D. Ayata, Y. Yaslan, and M. E. Kamasak, “Emotion based
music recommendation system using wearable physiological
were asked to report their mood in the pre-study and the sensors,” IEEE transactions on consumer electronics, vol. 64,
post study (Figure 3b). Participants were reported to have no. 2, pp. 196–203, 2018.
significantly higher level of valence (t = 2.244, p < 0.05) [4] T. Schäfer, P. Sedlmeier, C. Städtler, and D. Huron, “The
and arousal (t = 4.284, p < 0.01) after their experience psychological functions of music listening,” Frontiers in psy-
on EmoMusic. It should be mentioned that no participant chology, vol. 4, p. 511, 2013.
[5] Y. Song, “The role of emotion and context in musical prefer-
chose ‘falling asleep’ during this study. In addition, we can ence.” Ph.D. dissertation, Queen Mary University of London,
see in Figure 3c that the median level of satisfaction and 2016.
difficulty of using EmoMusic were 4 and 2 respectively, [6] M. H. Immordino-Yang, X.-F. Yang, and H. Damasio, “Corre-
which revealed good usability of the service. lations between social-emotional feelings and anterior insula
At the end of the study, participants were asked to give activity are independent from visceral states but influenced
by culture,” Frontiers in human neuroscience, vol. 8, p. 728,
comments on their experience and leave suggestions for 2014.
improving the service. On the positive side, many partic- [7] Y. Miyamoto, X. Ma, and A. G. Petermann, “Cultural differ-
ipants said the interface was easy to use (e.g, ‘nice and ences in hedonic emotion regulation after a negative event.”
clean’, ‘intuitive’) and they enjoyed the service. On the other Emotion, vol. 14, no. 4, p. 804, 2014.
hand, participants also mentioned that the music collection [8] M. Joshanloo, Z. K. Lepshokova, T. Panyusheva, A. Natalia,
W.-C. Poon, V. W.-l. Yeung, S. Sundaram, M. Achoui,
was small and lacked music of other genres they like such R. Asano, T. Igarashi et al., “Cross-cultural validation of
as ‘instrumental’. This problem can be addressed in the fear of happiness scale across 14 national groups,” Journal
future by extending the music collection with tracks of of Cross-Cultural Psychology, vol. 45, no. 2, pp. 246–264,
more genres and incorporating collaborative or content based 2014.
recommendation methods for personalization. [9] D. Müllensiefen, B. Gingras, J. Musil, and L. Stewart, “The
musicality of non-musicians: an index for assessing musical
IV. C ONCLUSION AND F UTURE W ORK sophistication in the general population.” PloS one, 2014.
[10] J. Posner, J. A. Russell, and B. S. Peterson, “The circum-
In this work, we investigated the emotional preference plex model of affect: An integrative approach to affective
of music in daily activities for context-aware music rec- neuroscience, cognitive development, and psychopathology,”
ommendation through a cross-cultural survey. Based on the Development and psychopathology, vol. 17, no. 3, pp. 715–
734, 2005.
emotional map between activity and music acquired from
[11] S. Stober and A. Nürnberger, “Adaptive music retrieval–a
the survey, we presented EmoMusic - a novel emotion-based state of the art,” Multimedia Tools and Applications, vol. 65,
music recommendation service for daily activities which no. 3, pp. 467–494, 2013.
supports explanation and control of music emotion through [12] K. Zhang, H. Zhang, S. Li, C. Yang, and L. Sun, “The pmemo
an interactive interface. A user study was presented to eval- dataset for music emotion recognition,” in Proceedings of
the 2018 acm on international conference on multimedia
uate the EmoMusic application and reported good feedback
retrieval, 2018, pp. 135–142.
from participants. In future work we will (i) explore further

353

Authorized licensed use limited to: University of Prince Edward Island. Downloaded on May 26,2021 at 11:43:16 UTC from IEEE Xplore. Restrictions apply.

You might also like