Bispectral features and mean shift clustering for stress and emotion recognition from natural speech
dc.authorid | 0000-0002-8929-3473 | en_US |
dc.authorid | 0000-0003-4526-0749 | en_US |
dc.authorid | 0000-0001-7466-0368 | en_US |
dc.authorid | 0000-0003-1840-9958 | en_US |
dc.contributor.author | C. K., Yogesh | |
dc.contributor.author | Muthusamy, Hariharan | |
dc.contributor.author | Rajamanickam, Yuvaraj | |
dc.contributor.author | Ngadiran, Ruzelita | |
dc.contributor.author | Adom, A. H. | |
dc.contributor.author | Polat, Kemal | |
dc.date.accessioned | 2021-06-23T19:45:36Z | |
dc.date.available | 2021-06-23T19:45:36Z | |
dc.date.issued | 2017 | |
dc.department | BAİBÜ, Mühendislik Fakültesi, Elektrik Elektronik Mühendisliği Bölümü | en_US |
dc.description.abstract | A new set of features and feature enhancement techniques are proposed to recognize emotion and stress from speech signal. The speech waveforms and the glottal waveforms (derived from the recorded emotional/stress speech waveforms) were processed by using third order statistics called bispectrum and 28 (14 from speech waveforms and 14 from glottal waveforms) bispectral based features. In this work, mean shift clustering was used to enhance the discrimination ability of the extracted Bispectral Features (BSFs). Four classifiers were used to distinguish different emotional and stressed states. The performance of the proposed method is tested with three databases. Different experiments were conducted and recognition rates were achieved in the range between 93.44% and 100% for Berlin emotional speech database (BES), between 73.81% and 97.23% for Surrey audio-visual expressed emotion database (SAVEE), between 93.8% and 100% for speech under simulated and actual stress simulated domain (SUSAS) (recognition of multi-style speech under stress-neutral, loud, lombard and anger) and 100% for SUSAS actual domain (recognition of three different levels of stress. high, medium and low). The obtained results indicate that the proposed bispectral based features and mean shift clustering provide promising results to recognize emotion and stress from speech signal. (C) 2017 Elsevier Ltd. All rights reserved. | en_US |
dc.identifier.doi | 10.1016/j.compeleceng.2017.01.024 | |
dc.identifier.endpage | 691 | en_US |
dc.identifier.issn | 0045-7906 | |
dc.identifier.issn | 1879-0755 | |
dc.identifier.scopus | 2-s2.0-85010987097 | en_US |
dc.identifier.scopusquality | Q1 | en_US |
dc.identifier.startpage | 676 | en_US |
dc.identifier.uri | https://doi.org/10.1016/j.compeleceng.2017.01.024 | |
dc.identifier.uri | https://hdl.handle.net/20.500.12491/9180 | |
dc.identifier.volume | 62 | en_US |
dc.identifier.wos | WOS:000413881700051 | en_US |
dc.identifier.wosquality | Q2 | en_US |
dc.indekslendigikaynak | Web of Science | en_US |
dc.indekslendigikaynak | Scopus | en_US |
dc.institutionauthor | Polat, Kemal | |
dc.language.iso | en | en_US |
dc.publisher | Pergamon-Elsevier Science Ltd | en_US |
dc.relation.ispartof | Computers & Electrical Engineering | en_US |
dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
dc.rights | info:eu-repo/semantics/closedAccess | en_US |
dc.subject | Speech Signals | en_US |
dc.subject | Glottal Signals | en_US |
dc.subject | Emotions | en_US |
dc.subject | Feature Extraction and Emotion Recognition | en_US |
dc.title | Bispectral features and mean shift clustering for stress and emotion recognition from natural speech | en_US |
dc.type | Article | en_US |
Dosyalar
Orijinal paket
1 - 1 / 1
Küçük Resim Yok
- İsim:
- yogesh-ck.pdf
- Boyut:
- 5.74 MB
- Biçim:
- Adobe Portable Document Format
- Açıklama:
- Tam Metin/Full Text