Improved speaker-independent emotion recognition from speech using two-stage feature reduction

In the recent years, researchers are focusing to improve the accuracy of speech emotion recognition. Generally, high emotion recognition accuracies were obtained for two-class emotion recognition, but multi-class emotion recognition is still a challenging task.The main aim of this work is to propose...

Full description

Bibliographic Details
Main Authors: Mohd Nazid, Hasrul, Muthusamy, Hariharan, Vijean, Vikneswaran, Yaacob, Sazali
Format: Article
Language:English
Published: Universiti Utara Malaysia Press 2015
Subjects:
Online Access:https://repo.uum.edu.my/id/eprint/24081/1/JICT%2014%202015%2057%E2%80%9376.pdf
Description
Summary:In the recent years, researchers are focusing to improve the accuracy of speech emotion recognition. Generally, high emotion recognition accuracies were obtained for two-class emotion recognition, but multi-class emotion recognition is still a challenging task.The main aim of this work is to propose a two-stage feature reduction using Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA) for improving the accuracy of the speech emotion recognition (ER) system. Short-term speech features were extracted from the emotional speech signals. Experiments were carried out using four different supervised classifi ers with two different emotional speech databases. From the experimental results, it can be inferred that the proposed method provides better accuracies of 87.48% for speaker dependent (SD) and gender dependent (GD) ER experiment, 85.15% for speaker independent (SI) ER experiment, and 87.09% for gender independent (GI) experiment.