Automatic speech based emotion recognition using paralinguistics features

dc.contributor.authorHook, J.
dc.contributor.authorNoroozi, F.
dc.contributor.authorToygar, O.
dc.contributor.authorAnbarjafari, G.
dc.date.accessioned2026-02-06T18:27:08Z
dc.date.issued2019
dc.departmentDoğu Akdeniz Üniversitesi
dc.description.abstractAffective computing studies and develops systems capable of detecting humans affects. The search for universal well-performing features for speech-based emotion recognition is ongoing. In this paper, a small set of features with support vector machines as the classifier is evaluated on Surrey Audio-Visual Expressed Emotion database, Berlin Database of Emotional Speech, Polish Emotional Speech database and Serbian emotional speech database. It is shown that a set of 87 features can offer results on-par with state-of-the-art, yielding 80.21, 88.6, 75.42 and 93.41% average emotion recognition rate, respectively. In addition, an experiment is conducted to explore the significance of gender in emotion recognition using random forests. Two models, trained on the first and second database, respectively, and four speakers were used to determine the effects. It is seen that the feature set used in this work performs well for both male and female speakers, yielding approximately 27% average emotion recognition in both models. In addition, the emotions for female speakers were recognized 18% of the time in the first model and 29% in the second. A similar effect is seen with male speakers: the first model yields 36%, the second 28% a verage emotion recognition rate. This illustrates the relationship between the constitution of training data and emotion recognition accuracy.
dc.description.sponsorshipEstonian Research Council [PUT638]; Scientific and Technological Research Council of Turkey (TUBITAK) [Proje 1001 - 116E097]; Estonian Centre of Excellence in IT (EXCITE) - European Regional Development Fund; European Network on Integrating Vision and Language (iV&L Net) ICT COST Action [IC1307]
dc.description.sponsorshipThis work has been partially supported by Estonian Research Council Grants (PUT638), The Scientific and Technological Research Council of Turkey (TUBITAK) (Proje 1001 - 116E097), the Estonian Centre of Excellence in IT (EXCITE) funded by the European Regional Development Fund and the European Network on Integrating Vision and Language (iV&L Net) ICT COST Action IC1307.
dc.identifier.doi10.24425/bpasts.2019.129647
dc.identifier.endpage488
dc.identifier.issn0239-7528
dc.identifier.issn2300-1917
dc.identifier.issue3
dc.identifier.orcid0000-0001-8460-5717
dc.identifier.orcid0000-0001-7402-9058
dc.identifier.orcid0000-0002-4618-1375
dc.identifier.scopus2-s2.0-85069195853
dc.identifier.scopusqualityQ2
dc.identifier.startpage479
dc.identifier.urihttps://doi.org/10.24425/bpasts.2019.129647
dc.identifier.urihttps://hdl.handle.net/11129/10796
dc.identifier.volume67
dc.identifier.wosWOS:000473332000005
dc.identifier.wosqualityQ3
dc.indekslendigikaynakWeb of Science
dc.indekslendigikaynakScopus
dc.language.isoen
dc.publisherPolska Akad Nauk, Polish Acad Sci, Div Iv Technical Sciences Pas
dc.relation.ispartofBulletin of the Polish Academy of Sciences-Technical Sciences
dc.relation.publicationcategoryMakale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı
dc.rightsinfo:eu-repo/semantics/openAccess
dc.snmzKA_WoS_20260204
dc.subjectrandom forests
dc.subjectspeech emotion recognition
dc.subjectmachine learning
dc.subjectsupport vector machines
dc.titleAutomatic speech based emotion recognition using paralinguistics features
dc.typeArticle

Files