Browse > Article
http://dx.doi.org/10.13088/jiis.2014.20.1.001

Multimodal Emotional State Estimation Model for Implementation of Intelligent Exhibition Services  

Lee, Kichun (Department of Industrial Engineering, Hanyang University)
Choi, So Yun (Graduate School of Business IT, Kookmin University)
Kim, Jae Kyeong (Graduate School of Business Administration, Kyung Hee University)
Ahn, Hyunchul (School of MIS, Kookmin University)
Publication Information
Journal of Intelligence and Information Systems / v.20, no.1, 2014 , pp. 1-14 More about this Journal
Abstract
Both researchers and practitioners are showing an increased interested in interactive exhibition services. Interactive exhibition services are designed to directly respond to visitor responses in real time, so as to fully engage visitors' interest and enhance their satisfaction. In order to install an effective interactive exhibition service, it is essential to adopt intelligent technologies that enable accurate estimation of a visitor's emotional state from responses to exhibited stimulus. Studies undertaken so far have attempted to estimate the human emotional state, most of them doing so by gauging either facial expressions or audio responses. However, the most recent research suggests that, a multimodal approach that uses people's multiple responses simultaneously may lead to better estimation. Given this context, we propose a new multimodal emotional state estimation model that uses various responses including facial expressions, gestures, and movements measured by the Microsoft Kinect Sensor. In order to effectively handle a large amount of sensory data, we propose to use stratified sampling-based MRA (multiple regression analysis) as our estimation method. To validate the usefulness of the proposed model, we collected 602,599 responses and emotional state data with 274 variables from 15 people. When we applied our model to the data set, we found that our model estimated the levels of valence and arousal in the 10~15% error range. Since our proposed model is simple and stable, we expect that it will be applied not only in intelligent exhibition services, but also in other areas such as e-learning and personalized advertising.
Keywords
감정 상태 추정 모형;멀티모달 접근법;지능형 전시 서비스;키넥트 센서;계층화 샘플링;
Citations & Related Records
Times Cited By KSCI : 4  (Citation Analysis)
연도 인용수 순위
1 Ahn, H., "GA-optimized Support Vector Regression for Better Emotional State Estimation Model," Proceedings of the 5th International Conference on Internet (ICONI 2013), (2013), 333-335.
2 Box, G. E. P. and. D. R. Cox, "An analysis of transformations," Journal of the Royal Statistical Society, Series B, Vol. 26, No. 2(1964), 211-252.
3 Cowie, R., E. Douglas-Cowie, S. Savvidou, E. McMahon, M. Sawey, and M. Schroder, "Feeltrace: An instrument for recording perceived emotion in real time," Proceedings of ISCA Workshop on Speech and Emotion, (2000), 19-24.
4 Jung, M. K., J. K. Kim, Y. Ryu, and H. Ahn, "An intelligent determination model of user's emotion using sequential neural network," Proceedings of International Conference on Marketing Studies (ICMS 2012), (2012).
5 Ekman, P., and W. Friesen, Facial Action Coding System: A Technique for the Measurement of Facial Movement. Consulting Psychologists Press, Palo Alto, 1978.
6 Hussain, A., E. Cambria, T. Mazzocco, M. Grassi, Q.-F. Wang, and T. Durrani, "Towards IMACA: Intelligent Multimodal Affective Conversational Agent," Lecture Notes in Computer Science, Vol. 7663(2012), 656-663.
7 Jung, M.-K., and J.-K. Kim, "The Intelligent Determination Model of Audience Emotion for Implementing Personalized Exhibition,"Journal of Intelligence and Information Systems, Vol. 18, No. 1(2012), 39-57.
8 Kim, C. K., Studies of Exhibit's Theory and Method, National Science Museum, Academic Series 12, 1996.
9 Kim, S., E. Ryoo, M. K. Jung, J. K. . Kim, and H. Ahn, "Application of Support Vector Regression for Improving the Performance of the Emotion Prediction Model," Journal of Intelligence and Information Systems, Vol. 18, No. 3(2012), 185-202.
10 Ko, K.-E., D.-J. Shin, and K.-B. Sim, "Development of context awareness and service reasoning technique for handicapped people," Journal of Korean Institute of Intelligent Systems, Vol. 18, No. 4(2008), 512-517.   DOI   ScienceOn
11 Lange, K., Numerical Analysis for Statisticians, Springer, 2010.
12 Nicolaou, M., H. Gunes, and M. Pantic, "Continuous Prediction of Spontaneous Affect from Multiple Cues and Modalities in Valence-Arousal Space," IEEE Transactions on Affective Computing, Vol. 2, No. 2(2011), 92-105.   DOI   ScienceOn
13 Nicolle, J., V. Rapp, K. Bailly. L. Prevost, and M. Chetouani, "Robust Continuous Prediction of Human Emotions using Multiscale Dynamic Cues," Proceedings of the 14th ACM International Conference on Multimodal Interaction (ICMI '12), (2012), 501-508.
14 Quesenberry, C. P. and N. P. Jewell, "Regression analysis based on stratified samples," Biometrika, Vol. 73, No. 3(1986), 605-614.   DOI   ScienceOn
15 Russell, J. A. "A Circumplex Model of Affect," Journal of Personality and Social Psychology, Vol. 39, No. 6(1980), 1161-1178.   DOI
16 Ryoo, E. C., H. Ahn, and J. K. Kim, "The Audience Behavior-based Emotion Prediction Model for Personalized Service," Journal of Intelligence and Information Systems, Vol. 19, No. 2(2013), 73-85.
17 Wollmer, M., M. Kaiser, F. Eyben, F. Weninger, B. Schuller, and G. Rigoll, "Fully Automatic Audiovisual Emotion Recognition: Voice, Words, and the Face," Proceedings of 10. ITG Symposium, Speech Communication, (2012), 1-4.
18 Sayre, S., "Multimedia that matters: Gallery-based technology and the museum visitor," First Monday, Vol. 10, No. 5(2005), doi:10. 5210/fm.v10i5.1244.   DOI
19 Wollmer, M., A. Metallinou, F. Eyben, B. Schuller, and S. Narayanan, "Context-Sensitive Multimodal Emotion Recognition from Speech and Facial Expression using Bidirectional LSTM Modeling," Proceedings of INTERSPEECH 2010, (2010), 2362-2365.
20 Wikipedia, Kinect, 2014. Available at http://en.wikipedia.org/wiki/Kinect (Accessed on January 15, 2014).