Paper
13 April 2018 Speaker emotion recognition: from classical classifiers to deep neural networks
Eya Mezghani, Maha Charfeddine, Henri Nicolas, Chokri Ben Amar
Author Affiliations +
Proceedings Volume 10696, Tenth International Conference on Machine Vision (ICMV 2017); 106962M (2018) https://doi.org/10.1117/12.2309476
Event: Tenth International Conference on Machine Vision, 2017, Vienna, Austria
Abstract
Speaker emotion recognition is considered among the most challenging tasks in recent years. In fact, automatic systems for security, medicine or education can be improved when considering the speech affective state. In this paper, a twofold approach for speech emotion classification is proposed. At the first side, a relevant set of features is adopted, and then at the second one, numerous supervised training techniques, involving classic methods as well as deep learning, are experimented. Experimental results indicate that deep architecture can improve classification performance on two affective databases, the Berlin Dataset of Emotional Speech and the SAVEE Dataset Surrey Audio-Visual Expressed Emotion.
© (2018) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Eya Mezghani, Maha Charfeddine, Henri Nicolas, and Chokri Ben Amar "Speaker emotion recognition: from classical classifiers to deep neural networks", Proc. SPIE 10696, Tenth International Conference on Machine Vision (ICMV 2017), 106962M (13 April 2018); https://doi.org/10.1117/12.2309476
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Neural networks

Classification systems

Feature extraction

Machine learning

Back to Top