Browse > Article
http://dx.doi.org/10.13064/KSSS.2020.12.3.025

An analysis of emotional English utterances using the prosodic distance between emotional and neutral utterances  

Yi, So-Pae (Humphreys West Elementary School)
Publication Information
Phonetics and Speech Sciences / v.12, no.3, 2020 , pp. 25-32 More about this Journal
Abstract
An analysis of emotional English utterances with 7 emotions (calm, happy, sad, angry, fearful, disgust, surprised) was conducted using the measurement of prosodic distance between 672 emotional and 48 neutral utterances. Applying the technique proposed in the automatic evaluation model of English pronunciation to the present study on emotional utterances, Euclidean distance measurement of 3 prosodic elements such as F0, intensity and duration extracted from emotional and neutral utterances was utilized. This paper, furthermore, extended the analytical methods to include Euclidean distance normalization, z-score and z-score normalization resulting in 4 groups of measurement schemes (sqrF0, sqrINT, sqrDUR; norsqrF0, norsqrINT, norsqrDUR; sqrzF0, sqrzINT, sqrzDUR; norsqrzF0, norsqrzINT, norsqrzDUR). All of the results from perceptual analysis and acoustical analysis of emotional utteances consistently indicated the greater effectiveness of norsqrF0, norsqrINT and norsqrDUR, among 4 groups of measurement schemes, which normalized the Euclidean measurement. The greatest acoustical change of prosodic information influenced by emotion was shown in the values of F0 followed by duration and intensity in descending order according to the effect size based on the estimation of distance between emotional utterances and neutral counterparts. Tukey Post Hoc test revealed 4 homogeneous subsets (calm
Keywords
emotional utterance; prosody; euclidean distance; arousal; valence; stance;
Citations & Related Records
Times Cited By KSCI : 1  (Citation Analysis)
연도 인용수 순위
1 Bachorowski, J., & Owren, M. J. (1995). Vocal expression of emotion: Acoustic properties of speech are associated with emotional intensity and context. Psychological Science, 6(4), 219-224.   DOI
2 Boersma, P., & Weenink, D. (2020). Praat: Doing phonetics by computer (version 6.1.16) [Computer program]. Retrieved from https://www.praat.org/
3 Breazeal, C. (2003). Emotion and sociable humanoid robots. International Journal of Human-Computer Studies, 59(1), 119-155.   DOI
4 Cahn, J. (1990). Generating expression in synthesized speech (Technical report). Boston, MA: MIT Media Lab.
5 Carlson, R., Granstrom, B., & Nord, L. (1992, October). Experiments with emotive speech: Acted utterances and synthesized replicas. Proceedings of the International Conference on Spoken Language Processing (ICSLP-92) (pp. 671-674). Banff, AB, Canada.
6 Kitahara, Y., & Tohkura, Y. (1992). Prosodic control to express emotions for man-machine interaction. IEICE Transactions on Fundamentals of Electronics: Communications and Computer Sciences, 75(2), 155-163.
7 Paulmann, S. (2016). The neurocognition of prosody. In G. Hickok, S. Small (Eds.), Neurobiology of language (pp. 1109-1120). San Diego, CA: Elsevier.
8 Kitayama, S., & Ishii, K. (2002). Word and voice: spontaneous attention to emotional utterances in two languages. Cognition and Emotion, 16(1), 29-59.   DOI
9 Livingstone, S. R., & Russo, F. A. (2018). The Ryerson Audio- Visual Database of Emotional Speech and Song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American english. PLOS ONE, 13(5), e0196391.   DOI
10 Mozziconacci, S. (1998). Speech variability and emotion: Production and perception (Doctoral dissertation). Technical University of Eindhoven, The Netherlands.
11 Pell, M., Paulmann, M., Dara, S., Alasseri, A., & Kotzb, S. (2009). Factors in the recognition of vocally expressed emotions: A comparison of our languages. Journal of Phonetics, 37(4), 417-435.   DOI
12 Thompson, W. F., & Balkwill, L. L. (2009). Cross-cultural similarities and differences. In P. N. Juslin, J. A. Sloboda (Eds.), Handbook of music and emotion: Theory, research, applications, 1st Edn, 755-791. New York, NY: Oxford University Press.
13 Vroomen, J., Collier, R., & Mozziconacci, S. (1993)., September). Duration and intonation in emotional speech. Proceedings of the 3rd European Conference on Speech Communication and Technology. Eurospeech-93, 577-580. Berlin, Germany.
14 Williams, C. & Stevens, K. (1972). Emotions and speech: Some acoustical correlates. The Journal of the Acoustical Society of America, 52(4B), 1238-1250.   DOI
15 Yi, S. P. (2018). Study on pitch contour extracted from Korean emotional speech using momel, Journal of Language Sciences, 25(3), 191-209.   DOI
16 Yoon, K. (2013). A study on human evaluators using the evaluation model of english pronunciation. Phonetics and Speech Sciences, 5(4), 109-119.   DOI
17 Yoon, K. (2009a). Building a sentential model for automatic prosody evaluation. Phonetics and Speech Sciences, 1(4), 47-59.
18 Yoon, K. (2009b). Synthesis and evaluation of prosodically exaggerated utterances. Phonetics and Speech Sciences, 1(3), 73-85.