A Pattern Mining Approach for Improving Speech Emotion Recognition

dc.contributor.author Umut Avci
dc.date.accessioned 2025-10-06T17:49:49Z
dc.date.issued 2022
dc.description.abstract Speech-driven user interfaces are becoming more common in our lives. To interact with such systems naturally and effectively machines need to recognize the emotional states of users and respond to them accordingly. At the heart of the emotion recognition research done to this end lies the emotion representation that enables machines to learn and predict emotions. Speech emotion recognition studies use a wide range of low-to-high-level acoustic features for representation purposes such as LLDs their functionals and BoAW. In this paper we present a new method for extracting a novel set of high-level features for classifying emotions. For this purpose we (1) reduce the dimension of discrete-time speech signals (2) perform a quantization operation on the new signals and assign a distinct symbol to each quantization level (3) use the symbol sequences representing the signals to extract discriminative patterns that are capable of distinguishing different emotions from each other and (4) generate a separate set of features for each emotion from the extracted patterns. Experimental results show that pattern features outperform Energy Voicing MFCC Spectral and RASTA feature sets. We also demonstrate that combining the pattern-based features and the acoustic features further improves the classification performance. © 2022 Elsevier B.V. All rights reserved.
dc.identifier.doi 10.1142/S0218001422500458
dc.identifier.issn 02180014, 17936381
dc.identifier.issn 0218-0014
dc.identifier.issn 1793-6381
dc.identifier.uri https://www.scopus.com/inward/record.uri?eid=2-s2.0-85143671883&doi=10.1142%2FS0218001422500458&partnerID=40&md5=abb36b04c42c043c850fad3d60e2fbe5
dc.identifier.uri https://gcris.yasar.edu.tr/handle/123456789/8647
dc.language.iso English
dc.publisher World Scientific
dc.relation.ispartof International Journal of Pattern Recognition and Artificial Intelligence
dc.source International Journal of Pattern Recognition and Artificial Intelligence
dc.subject Feature Extraction, Pattern Mining, Speech Emotion Recognition, Classification (of Information), Data Mining, Quantization (signal), Speech Recognition, User Interfaces, Acoustic Features, Emotion Representation, Emotional State, Features Extraction, Functionals, Learn+, Low-to-high, Pattern Mining, Speech Emotion Recognition, Emotion Recognition
dc.subject Classification (of information), Data mining, Quantization (signal), Speech recognition, User interfaces, Acoustic features, Emotion representation, Emotional state, Features extraction, Functionals, Learn+, Low-to-high, Pattern mining, Speech emotion recognition, Emotion Recognition
dc.title A Pattern Mining Approach for Improving Speech Emotion Recognition
dc.type Article
dspace.entity.type Publication
gdc.bip.impulseclass C5
gdc.bip.influenceclass C5
gdc.bip.popularityclass C5
gdc.coar.type text::journal::journal article
gdc.collaboration.industrial false
gdc.description.volume 36
gdc.identifier.openalex W4303449672
gdc.index.type Scopus
gdc.oaire.diamondjournal false
gdc.oaire.impulse 1.0
gdc.oaire.influence 2.4108997E-9
gdc.oaire.isgreen false
gdc.oaire.popularity 2.396017E-9
gdc.oaire.publicfunded false
gdc.oaire.sciencefields 0202 electrical engineering, electronic engineering, information engineering
gdc.oaire.sciencefields 02 engineering and technology
gdc.openalex.collaboration National
gdc.openalex.fwci 0.185
gdc.openalex.normalizedpercentile 0.57
gdc.opencitations.count 1
gdc.plumx.mendeley 1
gdc.plumx.newscount 1
gdc.plumx.scopuscites 1
person.identifier.scopus-author-id Avci- Umut (35486827300)
publicationissue.issueNumber 14
publicationvolume.volumeNumber 36
relation.isOrgUnitOfPublication ac5ddece-c76d-476d-ab30-e4d3029dee37
relation.isOrgUnitOfPublication.latestForDiscovery ac5ddece-c76d-476d-ab30-e4d3029dee37

Files