Translator Disclaimer
9 January 2008 An audiovisual emotion recognition system
Author Affiliations +
Proceedings Volume 6794, ICMIT 2007: Mechatronics, MEMS, and Smart Materials; 679426 (2008)
Event: ICMIT 2007: Mechatronics, MEMS, and Smart Materials, 2007, Gifu, Japan
Human emotions could be expressed by many bio-symbols. Speech and facial expression are two of them. They are both regarded as emotional information which is playing an important role in human-computer interaction. Based on our previous studies on emotion recognition, an audiovisual emotion recognition system is developed and represented in this paper. The system is designed for real-time practice, and is guaranteed by some integrated modules. These modules include speech enhancement for eliminating noises, rapid face detection for locating face from background image, example based shape learning for facial feature alignment, and optical flow based tracking algorithm for facial feature tracking. It is known that irrelevant features and high dimensionality of the data can hurt the performance of classifier. Rough set-based feature selection is a good method for dimension reduction. So 13 speech features out of 37 ones and 10 facial features out of 33 ones are selected to represent emotional information, and 52 audiovisual features are selected due to the synchronization when speech and video fused together. The experiment results have demonstrated that this system performs well in real-time practice and has high recognition rate. Our results also show that the work in multimodules fused recognition will become the trend of emotion recognition in the future.
© (2008) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Yi Han, Guoyin Wang, Yong Yang, and Kun He "An audiovisual emotion recognition system", Proc. SPIE 6794, ICMIT 2007: Mechatronics, MEMS, and Smart Materials, 679426 (9 January 2008);

Back to Top