• Title/Summary/Keyword: emotion prediction

Search Result 81, Processing Time 0.024 seconds

The Audience Behavior-based Emotion Prediction Model for Personalized Service (고객 맞춤형 서비스를 위한 관객 행동 기반 감정예측모형)

  • Ryoo, Eun Chung;Ahn, Hyunchul;Kim, Jae Kyeong
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.2
    • /
    • pp.73-85
    • /
    • 2013
  • Nowadays, in today's information society, the importance of the knowledge service using the information to creative value is getting higher day by day. In addition, depending on the development of IT technology, it is ease to collect and use information. Also, many companies actively use customer information to marketing in a variety of industries. Into the 21st century, companies have been actively using the culture arts to manage corporate image and marketing closely linked to their commercial interests. But, it is difficult that companies attract or maintain consumer's interest through their technology. For that reason, it is trend to perform cultural activities for tool of differentiation over many firms. Many firms used the customer's experience to new marketing strategy in order to effectively respond to competitive market. Accordingly, it is emerging rapidly that the necessity of personalized service to provide a new experience for people based on the personal profile information that contains the characteristics of the individual. Like this, personalized service using customer's individual profile information such as language, symbols, behavior, and emotions is very important today. Through this, we will be able to judge interaction between people and content and to maximize customer's experience and satisfaction. There are various relative works provide customer-centered service. Specially, emotion recognition research is emerging recently. Existing researches experienced emotion recognition using mostly bio-signal. Most of researches are voice and face studies that have great emotional changes. However, there are several difficulties to predict people's emotion caused by limitation of equipment and service environments. So, in this paper, we develop emotion prediction model based on vision-based interface to overcome existing limitations. Emotion recognition research based on people's gesture and posture has been processed by several researchers. This paper developed a model that recognizes people's emotional states through body gesture and posture using difference image method. And we found optimization validation model for four kinds of emotions' prediction. A proposed model purposed to automatically determine and predict 4 human emotions (Sadness, Surprise, Joy, and Disgust). To build up the model, event booth was installed in the KOCCA's lobby and we provided some proper stimulative movie to collect their body gesture and posture as the change of emotions. And then, we extracted body movements using difference image method. And we revised people data to build proposed model through neural network. The proposed model for emotion prediction used 3 type time-frame sets (20 frames, 30 frames, and 40 frames). And then, we adopted the model which has best performance compared with other models.' Before build three kinds of models, the entire 97 data set were divided into three data sets of learning, test, and validation set. The proposed model for emotion prediction was constructed using artificial neural network. In this paper, we used the back-propagation algorithm as a learning method, and set learning rate to 10%, momentum rate to 10%. The sigmoid function was used as the transform function. And we designed a three-layer perceptron neural network with one hidden layer and four output nodes. Based on the test data set, the learning for this research model was stopped when it reaches 50000 after reaching the minimum error in order to explore the point of learning. We finally processed each model's accuracy and found best model to predict each emotions. The result showed prediction accuracy 100% from sadness, and 96% from joy prediction in 20 frames set model. And 88% from surprise, and 98% from disgust in 30 frames set model. The findings of our research are expected to be useful to provide effective algorithm for personalized service in various industries such as advertisement, exhibition, performance, etc.

Visualization using Emotion Information in Movie Script (영화 스크립트 내 감정 정보를 이용한 시각화)

  • Kim, Jinsu
    • Journal of the Korea Convergence Society
    • /
    • v.9 no.11
    • /
    • pp.69-74
    • /
    • 2018
  • Through the convergence of Internet technology and various information technologies, it is possible to collect and process vast amount of information and to exchange various knowledge according to user's personal preference. Especially, there is a tendency to prefer intimate contents connected with the user's preference through the flow of emotional changes contained in the movie media. Based on the information presented in the script, the user seeks to visualize the flow of the entire emotion, the flow of emotions in a specific scene, or a specific scene in order to understand it more quickly. In this paper, after obtaining the raw data from the movie web page, it transforms it into a standardized scenario format after refining process. After converting the refined data into an XML document to easily obtain various information, various sentences are predicted by inputting each paragraph into the emotion prediction system. We propose a system that can easily understand the change of the emotional state between the characters in the whole or a specific part of the various emotions required by the user by mixing the predicted emotions flow and the amount of information included in the script.

A Study on the Emotion State Classification using Multi-channel EEG (다중채널 뇌파를 이용한 감정상태 분류에 관한 연구)

  • Kang, Dong-Kee;Kim, Heung-Hwan;Kim, Dong-Jun;Lee, Byung-Chae;Ko, Han-Woo
    • Proceedings of the KIEE Conference
    • /
    • 2001.07d
    • /
    • pp.2815-2817
    • /
    • 2001
  • This study describes the emotion classification using two different feature extraction methods for four-channel EEG signals. One of the methods is linear prediction analysis based on AR model. Another method is cross-correlation coefficients on frequencies of ${\theta}$, ${\alpha}$, ${\beta}$ bands. Using the linear predictor coefficients and the cross-correlation coefficients of frequencies, the emotion classification test for four emotions, such as anger, sad, joy, and relaxation is performed with a neural network. Comparing the results of two methods, it seems that the linear predictor coefficients produce the better results than the cross-correlation coefficients of frequencies for-emotion classification.

  • PDF

Convergence Implementing Emotion Prediction Neural Network Based on Heart Rate Variability (HRV) (심박변이도를 이용한 인공신경망 기반 감정예측 모형에 관한 융복합 연구)

  • Park, Sung Soo;Lee, Kun Chang
    • Journal of the Korea Convergence Society
    • /
    • v.9 no.5
    • /
    • pp.33-41
    • /
    • 2018
  • The purpose of this study is to develop more accurate and robust emotion prediction neural network (EPNN) model by combining heart rate variability (HRV) and neural network. For the sake of improving the prediction performance more reliably, the proposed EPNN model is based on various types of activation functions like hyperbolic tangent, linear, and Gaussian functions, all of which are embedded in hidden nodes to improve its performance. In order to verify the validity of the proposed EPNN model, a number of HRV metrics were calculated from 20 valid and qualified participants whose emotions were induced by using money game. To add more rigor to the experiment, the participants' valence and arousal were checked and used as output node of the EPNN. The experiment results reveal that the F-Measure for Valence and Arousal is 80% and 95%, respectively, proving that the EPNN yields very robust and well-balanced performance. The EPNN performance was compared with competing models like neural network, logistic regression, support vector machine, and random forest. The EPNN was more accurate and reliable than those of the competing models. The results of this study can be effectively applied to many types of wearable computing devices when ubiquitous digital health environment becomes feasible and permeating into our everyday lives.

Prediction of a rectal temperature utilizing a thermal perception index (열감지지수를 활용한 신체온도의 예측)

  • ;Jerry D. Ramsey
    • Science of Emotion and Sensibility
    • /
    • v.1 no.2
    • /
    • pp.75-79
    • /
    • 1998
  • 이 논문은 신체온도를 직접 측정하지 않고서 신체온도를 예상하는 모델을 연구한 것이다. 열감지지수 (TPI)를 개발하여 환경으로부터 느끼는 체감온도와 몸의 내부온도인 직장온도(Trec)와 몸의 외부온도인 피부온도 (Tsdin)를 예측하도록 하였다. Kwon과 Ramsey의 개발모델을 Goldman 의 모델과 비교해본 결과 정확도에 통계적으로 유의한 차이가 없었다. 회귀분석과 경험을 토대로 만든 체감온도를 예측할 수 있는 손쉬운 Kwon의 열감지지수 (KTPI)도 제시하였다. 대부분의 사람들이 쉽게 예측할 수 있도록 측정 또는 사용가능한 몇 개의 환경변수로부터 쉽게 몸의 예산 내부온도와 외부온도를 계산할 수 있게 단순화하였다.

  • PDF

Prediction of a rectal temperature utilizing a thermal perception index.

  • Kwon, Young G.;Jerry D.Ramsey
    • Proceedings of the Korean Society for Emotion and Sensibility Conference
    • /
    • 1998.04a
    • /
    • pp.159-164
    • /
    • 1998
  • 이 논문은 신체온도를 직접 측정하지 않고서 신체온도를 예상하는 모델을 연구한 것이다. 열감지지수 (TPI)를 개발하여 환경으로부터 느끼는 체감온도와 몸의 내부온도인 직장온도(Tuec)와 몸의 외부온도인 피부온도 (Tskin)를 예측하도록 하였다. Kwon와 Ramsey의 개발모델을 Goldman의 모델과 비교해본 결과 정확도에 통계적으로 유의한 차이가 없었다. 회귀분석과 경험을 토대로 만든 체감온도를 예측할 수 있는 손쉬운 Kwon의 열감지수 (KTPI)도 제시하였다. 대부분의 사람들이 쉽게 예측할 수 있도록 측정 또는 사용가능한 몇 개의 환경변수로부터 몸의 예상 내부온도와 외부온도를 계산할 수 있게 단순화하였다.

  • PDF

Convolutional Neural Network Model Using Data Augmentation for Emotion AI-based Recommendation Systems

  • Ho-yeon Park;Kyoung-jae Kim
    • Journal of the Korea Society of Computer and Information
    • /
    • v.28 no.12
    • /
    • pp.57-66
    • /
    • 2023
  • In this study, we propose a novel research framework for the recommendation system that can estimate the user's emotional state and reflect it in the recommendation process by applying deep learning techniques and emotion AI (artificial intelligence). To this end, we build an emotion classification model that classifies each of the seven emotions of angry, disgust, fear, happy, sad, surprise, and neutral, respectively, and propose a model that can reflect this result in the recommendation process. However, in the general emotion classification data, the difference in distribution ratio between each label is large, so it may be difficult to expect generalized classification results. In this study, since the number of emotion data such as disgust in emotion image data is often insufficient, correction is made through augmentation. Lastly, we propose a method to reflect the emotion prediction model based on data through image augmentation in the recommendation systems.

Deep Learning-Based Speech Emotion Recognition Technology Using Voice Feature Filters (음성 특징 필터를 이용한 딥러닝 기반 음성 감정 인식 기술)

  • Shin Hyun Sam;Jun-Ki Hong
    • The Journal of Bigdata
    • /
    • v.8 no.2
    • /
    • pp.223-231
    • /
    • 2023
  • In this study, we propose a model that extracts and analyzes features from deep learning-based speech signals, generates filters, and utilizes these filters to recognize emotions in speech signals. We evaluate the performance of emotion recognition accuracy using the proposed model. According to the simulation results using the proposed model, the average emotion recognition accuracy of DNN and RNN was very similar, at 84.59% and 84.52%, respectively. However, we observed that the simulation time for DNN was approximately 44.5% shorter than that of RNN, enabling quicker emotion prediction.

Emotion Prediction from Natural Language Documents ith Emotion Network (감정망을 활용한 자연언어 문서 상의 감정예측)

  • Min, Hye-Jin;Park, Jong-C.
    • Annual Conference on Human and Language Technology
    • /
    • 2004.10d
    • /
    • pp.191-199
    • /
    • 2004
  • 본 논문에서는 텍스트에 나타난 감정상태를 인지하는 모델을 제안하고, 이러한 모델을 활용하여 현재문장에서 나타난 감정 및 이후에 나타나게 될 감정상태들을 예측하는 시스템에 대하여 다룬다. 사용자의 감정을 인지하고 이에 대한 자연스러운 메시지, 행동 등을 통해 인간과 상호작용 할 수 있는 컴퓨터시스템을 구현하기 위해서는 현재의 감정상태뿐만 아니라 사용자 개개인의 정보 및 시스템과 상호작용하고 있는 상황의 정보 등을 통해 이후에 사용자가 느낄 수 있는 감정을 예측할 수 있는 감정모델이 요구된다. 본 논문에서는 파악된 이전의 감정상태 및 실제 감정과 표현된 감정간의 관계, 그리고 감정에 영향을 미친 주변대상의 특징 및 감정경험자의 목표와 행동이 반영된 상태-전이형태의 감정모델인 감정망(Emotion Network)을 제안한다. 감정망은 각 감정을 나타내는 상태(state)와 연결된 상태들 간의 전이(transition), 그리고 전이가 발생하기 위한 조건(condition)으로 구성된다. 본 논문에서는 텍스트 형태의 상담예시에 감정망을 활용하여 문헌의 감정어휘에 의해 직접적으로 표출되지 않는 감정을 예측할 수 있음을 보인다.

  • PDF

Exploring the Relationships Between Emotions and State Motivation in a Video-based Learning Environment

  • YU, Jihyun;SHIN, Yunmi;KIM, Dasom;JO, Il-Hyun
    • Educational Technology International
    • /
    • v.18 no.2
    • /
    • pp.101-129
    • /
    • 2017
  • This study attempted to collect learners' emotion and state motivation, analyze their inner states, and measure state motivation using a non-self-reported survey. Emotions were measured by learning segment in detailed learning situations, and they were used to indicate total state motivation with prediction power. Emotion was also used to explain state motivation by learning segment. The purpose of this study was to overcome the limitations of video-based learning environments by verifying whether the emotions measured during individual learning segments can be used to indicate the learner's state motivation. Sixty-eight students participated in a 90-minute to measure their emotions and state motivation, and emotions showed a statistically significant relationship between total state motivation and motivation by learning segment. Although this result is not clear because this was an exploratory study, it is meaningful that this study showed the possibility that emotions during different learning segments can indicate state motivation.