You are on page 1of 7

Acoustic Contrast Between Neutral and

Angry Speech: Variation of Prosodic


Features in Algerian Dialect Speech and
German Speech
Advances in Multidisciplinary Medical Technologies ─ Engineering, Modeling and
Findings pp 41-52 | Cite as

F. Ykhlef (1) Email author (fykhlef@cdta.dz)View author's OrcID profile


(View OrcID profile)
D. Bouchaffra (1) View author's OrcID profile (View OrcID profile)

1. Division Architecture des Systèmes et Multimédia, Centre de Développement des


Technologies Avancées, , Algiers, Algeria

Conference paper
First Online: 08 November 2020

18 Downloads

Abstract

Monitoring emotions through speech is crucial in medical psychology in addition


to emotional health. Angry speech automatic detection can be expedient in several
healthcare applications, e.g., (i) the estimation of the level of stress and (ii) the
incorporation of intelligence to nursing care robots. Profound knowledge of the
linguistic and acoustic characteristics of emotional speech assist categorizing
angry talking. This research manuscript elaborates on the audile distinction
among neutral as well as hostile talking by (i) probing the variation of prosodic
features, such as pitch (F0), energy (E), and duration (D), in the Algerian dialect
(AD), while (ii) comparing it with the German language (GL). The authors
recommend a tactic to quantify the separation among irate and neutral states
exploring subsets of emotional speech corpora of AD and GL. The authors
identified a noteworthy dissimilarity between AD and GL regarding the deviations
of neutral and angry prosodic features.

Keywords
German language  Algerian dialect  Acoustic contrast  Angry speech 
Neutral speech  Emotion analysis  Automatic speech analysis 
This is a preview of subscription content, log in to check access.

Notes

Acknowledgments

This research was funded by the ATRSS (Agence Thématique de Recherche en


Science de la Santé) under grant n°: 64/DFPR/ATRSS/2017.

References
1. H. Kaya, F. Gürpinar, A.A. Salah, Video-based emotion recognition in the
wild using deep transfer learning and score fusion. Image Vis. Comput. 65,
66–75 (2017)
CrossRef (https://doi.org/10.1016/j.imavis.2017.01.012)
Google Scholar (http://scholar.google.com/scholar_lookup?title=Video-
based%20emotion%20recognition%20in%20the%20wild%20using%20deep
%20transfer%20learning%20and%20score%20fusion&author=H.%20Kaya&
author=F.%20G%C3%BCrpinar&author=AA.%20Salah&journal=Image%20Vi
s.%20Comput.&volume=65&pages=66-75&publication_year=2017)
2. H.R. Marins, V.V. Estrela, On the use of motion vectors for 2D and 3D error
concealment in H.264 AVC video, in Feature Detectors and Motion Detection
in Video Processing, ed. by N. Dey, A. S. Ashour, P. K. Patra, (2017).
https://doi.org/10.4018/978-1-5225-1025-3.ch008 (https://doi.org/10.4018/978-
1-5225-1025-3.ch008)
CrossRef (https://doi.org/10.4018/978-1-5225-1025-3.ch008)
Google Scholar (http://scholar.google.com/scholar_lookup?
title=On%20the%20use%20of%20motion%20vectors%20for%202D%20and%
203D%20error%20concealment%20in%20H.264%20AVC%20video&author=
HR.%20Marins&author=VV.%20Estrela&publication_year=2017)
3. A.E. Herrmann, V.V. Estrela, Content-based image retrieval (CBIR) in remote
clinical diagnosis and healthcare, in Encyclopedia of E-Health and
Telemedicine, ed. by M. M. Cruz-Cunha, I. M. Miranda, R. Martinho, R. Rijo,
(2016).  https://doi.org/10.4018/978-1-4666-9978-6.ch039
(https://doi.org/10.4018/978-1-4666-9978-6.ch039)
CrossRef (https://doi.org/10.4018/978-1-4666-9978-6.ch039)
Google Scholar (http://scholar.google.com/scholar_lookup?title=Content-
based%20image%20retrieval%20%28CBIR%29%20in%20remote%20clinical
%20diagnosis%20and%20healthcare&author=AE.%20Herrmann&author=V
V.%20Estrela&publication_year=2016)
4. C. Su, W. Zhang, T. Feng, Research on Extracting Facial Image for Bimodal
Emotion Recognition Based on Speech Signal and Facial Expression. Int. J.
Eng. Tech. 4(1), 589–594, (2018)
Google Scholar (https://scholar.google.com/scholar?
q=C.%20Su%2C%20W.%20Zhang%2C%20T.%20Feng%2C%20Research%20on
%20Extracting%20Facial%20Image%20for%20Bimodal%20Emotion%20Rec
ognition%20Based%20on%20Speech%20Signal%20and%20Facial%20Expres
sion.%20Int.%20J.%20Eng.%20Tech.%204%281%29%2C%20589%E2%80%93
594%2C%20%282018%29)
5. V.V. Estrela, A.M. Coelho, State-of-the-art motion estimation in the context of
3D TV, in Multimedia Networking and Coding, ed. by R. A. Farrugia, C. J.
Debono, (IGI Global, Hershey, 2013), pp. 148–173.
https://doi.org/10.4018/978-1-4666-2660-7.ch006 (https://doi.org/10.4018/978-
1-4666-2660-7.ch006)
CrossRef (https://doi.org/10.4018/978-1-4666-2660-7.ch006)
Google Scholar (http://scholar.google.com/scholar_lookup?title=State-of-the-
art%20motion%20estimation%20in%20the%20context%20of%203D%20TV&
author=VV.%20Estrela&author=AM.%20Coelho&pages=148-
173&publication_year=2013)
6. B. Sun, Q. Xu, J. He, L. Yu, L. Li, Q. Wei, Audio-Video Based Multimodal
Emotion Recognition Using SVMs and Deep Learning (CCPR, 2016)
Google Scholar (https://scholar.google.com/scholar?
q=B.%20Sun%2C%20Q.%20Xu%2C%20J.%20He%2C%20L.%20Yu%2C%20L.
%20Li%2C%20Q.%20Wei%2C%20Audio-
Video%20Based%20Multimodal%20Emotion%20Recognition%20Using%20S
VMs%20and%20Deep%20Learning%20%28CCPR%2C%202016%29)
7. J. Hook, F. Noroozi, Ö. Toygar, G. Anbarjafari, Automatic speech based
emotion recognition using paralinguistics features. Bull. Pol. Acad. Sci. Tech.
Sci. 67(3) (2019)
Google Scholar (https://scholar.google.com/scholar?
q=J.%20Hook%2C%20F.%20Noroozi%2C%20%C3%96.%20Toygar%2C%20G.
%20Anbarjafari%2C%20Automatic%20speech%20based%20emotion%20rec
ognition%20using%20paralinguistics%20features.%20Bull.%20Pol.%20Acad.
%20Sci.%20Tech.%20Sci.%2067%283%29%20%282019%29)
8. J.S. Adelman, Z. Estes, M. Cossu, Emotional sound symbolism: Languages
rapidly signal valence via phonemes. Cognition 175, 122–130 (2018)
CrossRef (https://doi.org/10.1016/j.cognition.2018.02.007)
Google Scholar (http://scholar.google.com/scholar_lookup?
title=Emotional%20sound%20symbolism%3A%20Languages%20rapidly%20
signal%20valence%20via%20phonemes&author=JS.%20Adelman&author=Z.
%20Estes&author=M.%20Cossu&journal=Cognition&volume=175&pages=12
2-130&publication_year=2018)
9. F. Vogt, Real-Time Automatic Emotion Recognition from Speech (PhD Thesis,
University of Bielefeld, Germany, 2010)
Google Scholar (https://scholar.google.com/scholar?
q=F.%20Vogt%2C%20Real-
Time%20Automatic%20Emotion%20Recognition%20from%20Speech%20%2
8PhD%20Thesis%2C%20University%20of%20Bielefeld%2C%20Germany%2C
%202010%29)
10. J. Deng, F. Eyben, B. Schuller, F. Burkhardt, in The 17th IEEE International
Conference on Affective Computing and Intelligent Interaction Workshops and
Demos, Deep neural networks for anger detection from real life speech data
(2017), pp. 1–6
Google Scholar (https://scholar.google.com/scholar?
q=J.%20Deng%2C%20F.%20Eyben%2C%20B.%20Schuller%2C%20F.%20Burk
hardt%2C%20in%20The%2017th%20IEEE%20International%20Conference
%20on%20Affective%20Computing%20and%20Intelligent%20Interaction%2
0Workshops%20and%20Demos%2C%20Deep%20neural%20networks%20fo
r%20anger%20detection%20from%20real%20life%20speech%20data%20%2
82017%29%2C%20pp.%201%E2%80%936)
11. F. Burkhardt, A. Paeschke, M. Rolfes, W.F. Sendlmeier, B. Weiss, in The 9th
European Conference on Speech Communication and Technology. A database
of German emotional speech (Lisboa, 2005)
Google Scholar (https://scholar.google.com/scholar?
q=F.%20Burkhardt%2C%20A.%20Paeschke%2C%20M.%20Rolfes%2C%20W.
F.%20Sendlmeier%2C%20B.%20Weiss%2C%20in%20The%209th%20Europe
an%20Conference%20on%20Speech%20Communication%20and%20Techno
logy.%20A%20database%20of%20German%20emotional%20speech%20%28
Lisboa%2C%202005%29)
12. M. Oliveira Jr, A.N. de Almeida, R.A. de Almeida, E.W. Silva, in The 7th
International Conference on Speech Prosody. Speech rate in the expression of
anger: a study with spontaneous speech material. (Dublin, 2014)
Google Scholar (https://scholar.google.com/scholar?
q=M.%20Oliveira%20Jr%2C%20A.N.%20de%20Almeida%2C%20R.A.%20de%
20Almeida%2C%20E.W.%20Silva%2C%20in%20The%207th%20International
%20Conference%20on%20Speech%20Prosody.%20Speech%20rate%20in%20
the%20expression%20of%20anger%3A%20a%20study%20with%20spontane
ous%20speech%20material.%20%28Dublin%2C%202014%29)
13. S.G. Koolagudi, S. Ray, K.S. Rao, in The International Conference on
Contemporary Computing. Emotion classification based on speaking rate
(India, 2010) pp. 316–327
Google Scholar (https://scholar.google.com/scholar?
q=S.G.%20Koolagudi%2C%20S.%20Ray%2C%20K.S.%20Rao%2C%20in%20Th
e%20International%20Conference%20on%20Contemporary%20Computing.
%20Emotion%20classification%20based%20on%20speaking%20rate%20%2
8India%2C%202010%29%20pp.%20316%E2%80%93327)
14. S. Yildirim, M. Bulut, C.M. Lee, A. Kazemzadeh, Z. Deng, S. Lee, C. Busso, in
The 8th International Conference on Spoken Language Processing. An
acoustic study of emotions expressed in speech (South Korea, 2014)
Google Scholar (https://scholar.google.com/scholar?
q=S.%20Yildirim%2C%20M.%20Bulut%2C%20C.M.%20Lee%2C%20A.%20Ka
zemzadeh%2C%20Z.%20Deng%2C%20S.%20Lee%2C%20C.%20Busso%2C%2
0in%20The%208th%20International%20Conference%20on%20Spoken%20L
anguage%20Processing.%20An%20acoustic%20study%20of%20emotions%2
0expressed%20in%20speech%20%28South%20Korea%2C%202014%29)
15. J.D. Gibbons, S. Chakraborti, Nonparametric Statistical Inference, 5th edn.
(Chapman & Hall/CRC Press/Taylor & Francis Group, Boca Raton, 2011)
zbMATH (http://www.emis.de/MATH-item?1278.62004)
Google Scholar (http://scholar.google.com/scholar_lookup?
title=Nonparametric%20Statistical%20Inference&author=JD.%20Gibbons&a
uthor=S.%20Chakraborti&publication_year=2011)
16. S. Harrat, K. Meftouh, M. Abbas, W.K. Hidouci, K. Smaili, An Algerian dialect:
Study and resources. Int. J. Adv. Comput. Sci. Appl. 7(3), 384–396 (2016)
Google Scholar (http://scholar.google.com/scholar_lookup?
title=An%20Algerian%20dialect%3A%20Study%20and%20resources&autho
r=S.%20Harrat&author=K.%20Meftouh&author=M.%20Abbas&author=WK.
%20Hidouci&author=K.%20Smaili&journal=Int.%20J.%20Adv.%20Comput.%
20Sci.%20Appl.&volume=7&issue=3&pages=384-
396&publication_year=2016)
17. U. Ammon, The Position of the German Language in the World (Taylor &
Francis Group, 2019)
Google Scholar (https://scholar.google.com/scholar?
q=U.%20Ammon%2C%20The%20Position%20of%20the%20German%20Lan
guage%20in%20the%20World%20%28Taylor%20%26%20Francis%20Group
%2C%202019%29)
18. F. Ykhlef, A. Derbal, W. Benzaba, R. Boutaleb, F. Ykhlef, D. Bouchaffra, H.
Meraoubi, in The IEEE International Conference on Advanced Electrical
Engineering. Towards building an emotional speech corpus of Algerian
dialect: criteria and preliminary assessment results. (Algiers, 2019)
Google Scholar (https://scholar.google.com/scholar?
q=F.%20Ykhlef%2C%20A.%20Derbal%2C%20W.%20Benzaba%2C%20R.%20
Boutaleb%2C%20F.%20Ykhlef%2C%20D.%20Bouchaffra%2C%20H.%20Mera
oubi%2C%20in%20The%20IEEE%20International%20Conference%20on%20
Advanced%20Electrical%20Engineering.%20Towards%20building%20an%2
0emotional%20speech%20corpus%20of%20Algerian%20dialect%3A%20crit
eria%20and%20preliminary%20assessment%20results.%20%28Algiers%2C
%202019%29)
19. F. Burkhardt, in The 18th International Conference on Language Resources
and Evaluation. “You Seem Aggressive!” Monitoring anger in a practical
application (Istanbul, 2012), pp. 1221–1225
Google Scholar (https://scholar.google.com/scholar?
q=F.%20Burkhardt%2C%20in%20The%2018th%20International%20Confere
nce%20on%20Language%20Resources%20and%20Evaluation.%20%E2%80
%9CYou%20Seem%20Aggressive%21%E2%80%9D%20Monitoring%20anger
%20in%20a%20practical%20application%20%28Istanbul%2C%202012%29%
2C%20pp.%201221%E2%80%931225)
20. S. Gonzalez, M. Brookes, PEFAC – a pitch estimation algorithm robust to high
levels of noise. IEEE Trans. Audio Speech Lang. Process. 22(2), 518–530
(2014)
CrossRef (https://doi.org/10.1109/TASLP.2013.2295918)
Google Scholar (http://scholar.google.com/scholar_lookup?
title=PEFAC%C2%A0%E2%80%93%20a%20pitch%20estimation%20algorith
m%20robust%20to%20high%20levels%20of%20noise&author=S.%20Gonzal
ez&author=M.%20Brookes&journal=IEEE%20Trans.%20Audio%20Speech%2
0Lang.%20Process.&volume=22&issue=2&pages=518-
530&publication_year=2014)
21. R.C. Guido, Enhancing teager energy operator based on a novel and
appealing concept: signal mass. J. Franklin Inst. 356(4), 2346–2352 (2019)
MathSciNet (http://www.ams.org/mathscinet-getitem?mr=3925989)
CrossRef (https://doi.org/10.1016/j.jfranklin.2018.12.007)
Google Scholar (http://scholar.google.com/scholar_lookup?
title=Enhancing%20teager%20energy%20operator%20based%20on%20a%2
0novel%20and%20appealing%20concept%3A%20signal%20mass&author=R
C.%20Guido&journal=J.%20Franklin%20Inst.&volume=356&issue=4&pages=
2346-2352&publication_year=2019)
22. T.W. Anderson, D.A. Darling, Asymptotic theory of certain “goodness-of-fit”
criteria based on stochastic processes. Ann. Math. Stat. 23, 193–212 (1952)
MathSciNet (http://www.ams.org/mathscinet-getitem?mr=50238)
CrossRef (https://doi.org/10.1214/aoms/1177729437)
Google Scholar (http://scholar.google.com/scholar_lookup?
title=Asymptotic%20theory%20of%20certain%20%E2%80%9Cgoodness-of-
fit%E2%80%9D%20criteria%20based%20on%20stochastic%20processes&au
thor=TW.%20Anderson&author=DA.%20Darling&journal=Ann.%20Math.%2
0Stat.&volume=23&pages=193-212&publication_year=1952)
23. I.C. Marschner, Inference Principles for Biostatisticians, 1st edn. (Chapman &
Hall/CRC Press/Taylor & Francis Group, Boca Raton, 2014)
CrossRef (https://doi.org/10.1201/b17824)
Google Scholar (http://scholar.google.com/scholar_lookup?
title=Inference%20Principles%20for%20Biostatisticians&author=IC.%20Mar
schner&publication_year=2014)

Copyright information
© Springer Nature Switzerland AG 2021

About this paper


Cite this paper as:
Ykhlef F., Bouchaffra D. (2021) Acoustic Contrast Between Neutral and Angry Speech: Variation of
Prosodic Features in Algerian Dialect Speech and German Speech. In: Khelassi A., Estrela V.V. (eds)
Advances in Multidisciplinary Medical Technologies ─ Engineering, Modeling and Findings. Springer,
Cham. https://doi.org/10.1007/978-3-030-57552-6_4

First Online 08 November 2020


DOI https://doi.org/10.1007/978-3-030-57552-6_4
Publisher Name Springer, Cham
Print ISBN 978-3-030-57551-9
Online ISBN 978-3-030-57552-6
eBook Packages Engineering Engineering (R0)

Buy this book on publisher's site


Reprints and Permissions

Personalised recommendations
© 2020 Springer Nature Switzerland AG. Part of Springer Nature.

Not logged in Not affiliated 179.187.226.25

You might also like