Browse > Article

A Study on The Improvement of Emotion Recognition by Gender Discrimination  

Cho, Youn-Ho (Dept. of Computer Science & Engineering, Dankook University)
Park, Kyu-Sik (Dept. of Computer Science & Engineering, Dankook University)
Publication Information
Abstract
In this paper, we constructed a speech emotion recognition system that classifies four emotions - neutral, happy, sad, and anger from speech based on male/female gender discrimination. At first, the proposed system distinguish between male and female from a queried speech, then the system performance can be improved by using separate optimized feature vectors for each gender for the emotion classification. As a emotion feature vector, this paper adopts ZCPA(Zero Crossings with Peak Amplitudes) which is well known for its noise-robustic characteristic from the speech recognition area and the features are optimized using SFS method. For a pattern classification of emotion, k-NN and SVM classifiers are compared experimentally. From the computer simulation results, the proposed system was proven to be highly efficient for speech emotion classification about 85.3% regarding four emotion states. This might promise the use the proposed system in various applications such as call-center, humanoid robots, ubiquitous, and etc.
Keywords
ZCPA; SVM; SFS;
Citations & Related Records
연도 인용수 순위
  • Reference
1 강봉석, "음성 신호를 이용한 문장독립 감정 인식 시스템," 석사학위 논문, 연세대학교, 2001
2 Dellaert, F., Polzin, T., and Waibel, A., "Recognizing Emotion in Speech," Proceedings of the International Conference on Spoken Language, Vol. 3, pp. 1970-1973, Oct. 1996
3 Vogt T and Andre E., "Improving Automatic Emotion Recognition from Speech via Gender Differentiation," Proceedings of Language Resources and Evaluation Conference '06, Italy, May 2006
4 Duda R., Hart P. and Stork D., "Pattern Classification, Second Edition," John Wiley & Sons, 2000
5 Jain A. and Zongker D., "Feature Selection: Evaluation, Application, and Small Sample Performance," IEEE transactions on pattern analysis and machine intelligence, Vol. 19, No. 2, pp.153-158. 1997   DOI   ScienceOn
6 Lugger M. and Yang B., "The relevance of voice quality features in speaker independent emotion recognition," Acoustics, Speech and Signal Processing, ICASSP 07., Vol. 4, pp.IV-17-IV-20, April 2007
7 Lee C. M. and Narayanan S. S., "Towards Detecting Emotions in Spoken Dialogs," IEEE Transactions on Speech and Audio Processing, Vol. 13, No. 2, March 2005
8 Doh-Suk Kim, Soo-Young Lee and Rhee M. Kil, "Auditory Processing of Speech Signals for Robust Speech Recognition in Real-World Noisy Environments," IEEE Transactions on Speech and Audio Processing, Vol. 7, No. 1, pp.55-69, Jan. 1999   DOI   ScienceOn
9 Xuejing Sun, "A Pitch Determination Algorithm Based On Subharmonic-to-Harmonic Ratio," International Conference on Spoken Language Processing '2000, pp.676-679, 2000
10 Lingyun Gu and Stephen A. Zahorian, "A New Robust Algorithm for Isolated Word Endpoint Detection," Acoustics, Speech and Signal Processing, ICASSP 02., Orlando, FL, May 2002
11 Rong J., Chen Y., Chowdhury M. and Li G., "Acoustic Features Extraction for Emotion Recognition," Computer and Information Science, ICIS 2007. 6th IEEE/ACIS International Conference, pp.419-424, July 2007
12 Dellaert, F., Polzin, T., and Waibel, A., "Recognizing Emotion in Speech," Proceedings of the International Conference on Spoken Language, Vol. 3, pp. 1970-1973, Oct. 1996