Automatic speech based emotion recognition using paralinguistics features
| dc.contributor.author | Hook, J. | |
| dc.contributor.author | Noroozi, F. | |
| dc.contributor.author | Toygar, O. | |
| dc.contributor.author | Anbarjafari, G. | |
| dc.date.accessioned | 2026-02-06T18:27:08Z | |
| dc.date.issued | 2019 | |
| dc.department | Doğu Akdeniz Üniversitesi | |
| dc.description.abstract | Affective computing studies and develops systems capable of detecting humans affects. The search for universal well-performing features for speech-based emotion recognition is ongoing. In this paper, a small set of features with support vector machines as the classifier is evaluated on Surrey Audio-Visual Expressed Emotion database, Berlin Database of Emotional Speech, Polish Emotional Speech database and Serbian emotional speech database. It is shown that a set of 87 features can offer results on-par with state-of-the-art, yielding 80.21, 88.6, 75.42 and 93.41% average emotion recognition rate, respectively. In addition, an experiment is conducted to explore the significance of gender in emotion recognition using random forests. Two models, trained on the first and second database, respectively, and four speakers were used to determine the effects. It is seen that the feature set used in this work performs well for both male and female speakers, yielding approximately 27% average emotion recognition in both models. In addition, the emotions for female speakers were recognized 18% of the time in the first model and 29% in the second. A similar effect is seen with male speakers: the first model yields 36%, the second 28% a verage emotion recognition rate. This illustrates the relationship between the constitution of training data and emotion recognition accuracy. | |
| dc.description.sponsorship | Estonian Research Council [PUT638]; Scientific and Technological Research Council of Turkey (TUBITAK) [Proje 1001 - 116E097]; Estonian Centre of Excellence in IT (EXCITE) - European Regional Development Fund; European Network on Integrating Vision and Language (iV&L Net) ICT COST Action [IC1307] | |
| dc.description.sponsorship | This work has been partially supported by Estonian Research Council Grants (PUT638), The Scientific and Technological Research Council of Turkey (TUBITAK) (Proje 1001 - 116E097), the Estonian Centre of Excellence in IT (EXCITE) funded by the European Regional Development Fund and the European Network on Integrating Vision and Language (iV&L Net) ICT COST Action IC1307. | |
| dc.identifier.doi | 10.24425/bpasts.2019.129647 | |
| dc.identifier.endpage | 488 | |
| dc.identifier.issn | 0239-7528 | |
| dc.identifier.issn | 2300-1917 | |
| dc.identifier.issue | 3 | |
| dc.identifier.orcid | 0000-0001-8460-5717 | |
| dc.identifier.orcid | 0000-0001-7402-9058 | |
| dc.identifier.orcid | 0000-0002-4618-1375 | |
| dc.identifier.scopus | 2-s2.0-85069195853 | |
| dc.identifier.scopusquality | Q2 | |
| dc.identifier.startpage | 479 | |
| dc.identifier.uri | https://doi.org/10.24425/bpasts.2019.129647 | |
| dc.identifier.uri | https://hdl.handle.net/11129/10796 | |
| dc.identifier.volume | 67 | |
| dc.identifier.wos | WOS:000473332000005 | |
| dc.identifier.wosquality | Q3 | |
| dc.indekslendigikaynak | Web of Science | |
| dc.indekslendigikaynak | Scopus | |
| dc.language.iso | en | |
| dc.publisher | Polska Akad Nauk, Polish Acad Sci, Div Iv Technical Sciences Pas | |
| dc.relation.ispartof | Bulletin of the Polish Academy of Sciences-Technical Sciences | |
| dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | |
| dc.rights | info:eu-repo/semantics/openAccess | |
| dc.snmz | KA_WoS_20260204 | |
| dc.subject | random forests | |
| dc.subject | speech emotion recognition | |
| dc.subject | machine learning | |
| dc.subject | support vector machines | |
| dc.title | Automatic speech based emotion recognition using paralinguistics features | |
| dc.type | Article |










