Emotions recognition: different sets of features and models
Autor: | A. Revathi, C. Jeyalakshmi |
---|---|
Rok vydání: | 2018 |
Předmět: |
Linguistics and Language
Computer science Speech recognition Language and Linguistics Human-Computer Interaction Identification (information) ComputingMethodologies_PATTERNRECOGNITION Formant Cepstrum Feature (machine learning) Computer Vision and Pattern Recognition Mel-frequency cepstrum Cluster analysis Affective computing Hidden Markov model Software |
Zdroj: | International Journal of Speech Technology. 22:473-482 |
ISSN: | 1572-8110 1381-2416 |
DOI: | 10.1007/s10772-018-9533-6 |
Popis: | The better and effective human–machine communication is ensured by performing affective computing. In the recent years, healthy research has been progressing in recognizing emotions by using various databases. This paper mainly emphasizes the effectiveness on the basis of using different sets of features and modeling techniques in evaluating the performance of multiple speaker-independent and speaker-dependent emotion recognition systems. It has become a challenging task to improve the performance of emotion recognition system, since EMO-DB Berlin database used in this work contains only ten speeches uttered by ten speakers in different emotions namely, Anger, Boredom, Disgust, Fear, Happiness, Sadness and Neutral. Speaker dependent and independent emotion recognition is done by creating models using clustering technique, Gaussian mixture modeling (GMM) and continuous density hidden Markov modeling (CDHMM) techniques for all emotions. The emotion recognition system is also evaluated for mel frequency cepstrum (MFCC) and concatenated MFCC with probability & shifted delta cepstrum (SDC), mel frequency linear predictive cepstrum (MFPLPC) and concatenated MFPLPC with probability & SDC and formants for clustering used as a modeling technique. These features provide complementary evidence in assessing the performance of the system based on VQ based clustering technique. This algorithm provides 99 and 100% as overall weighted accuracy recall (WAR) for performance evaluation with respect to correct identification of emotion for any one feature and modeling technique. |
Databáze: | OpenAIRE |
Externí odkaz: |