• Title/Summary/Keyword: frequency of emotion

Search Result 347, Processing Time 0.026 seconds

Voice Frequency Synthesis using VAW-GAN based Amplitude Scaling for Emotion Transformation

  • Kwon, Hye-Jeong;Kim, Min-Jeong;Baek, Ji-Won;Chung, Kyungyong
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.2
    • /
    • pp.713-725
    • /
    • 2022
  • Mostly, artificial intelligence does not show any definite change in emotions. For this reason, it is hard to demonstrate empathy in communication with humans. If frequency modification is applied to neutral emotions, or if a different emotional frequency is added to them, it is possible to develop artificial intelligence with emotions. This study proposes the emotion conversion using the Generative Adversarial Network (GAN) based voice frequency synthesis. The proposed method extracts a frequency from speech data of twenty-four actors and actresses. In other words, it extracts voice features of their different emotions, preserves linguistic features, and converts emotions only. After that, it generates a frequency in variational auto-encoding Wasserstein generative adversarial network (VAW-GAN) in order to make prosody and preserve linguistic information. That makes it possible to learn speech features in parallel. Finally, it corrects a frequency by employing Amplitude Scaling. With the use of the spectral conversion of logarithmic scale, it is converted into a frequency in consideration of human hearing features. Accordingly, the proposed technique provides the emotion conversion of speeches in order to express emotions in line with artificially generated voices or speeches.

Speech Emotion Recognition Using 2D-CNN with Mel-Frequency Cepstrum Coefficients

  • Eom, Youngsik;Bang, Junseong
    • Journal of information and communication convergence engineering
    • /
    • v.19 no.3
    • /
    • pp.148-154
    • /
    • 2021
  • With the advent of context-aware computing, many attempts were made to understand emotions. Among these various attempts, Speech Emotion Recognition (SER) is a method of recognizing the speaker's emotions through speech information. The SER is successful in selecting distinctive 'features' and 'classifying' them in an appropriate way. In this paper, the performances of SER using neural network models (e.g., fully connected network (FCN), convolutional neural network (CNN)) with Mel-Frequency Cepstral Coefficients (MFCC) are examined in terms of the accuracy and distribution of emotion recognition. For Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS) dataset, by tuning model parameters, a two-dimensional Convolutional Neural Network (2D-CNN) model with MFCC showed the best performance with an average accuracy of 88.54% for 5 emotions, anger, happiness, calm, fear, and sadness, of men and women. In addition, by examining the distribution of emotion recognition accuracies for neural network models, the 2D-CNN with MFCC can expect an overall accuracy of 75% or more.

1/f-LIKE FREQUENCY FLUCTUATION IN FRONTAL ALPHA WAVE AS AN INDICATOR OF EMOTION

  • Yoshida, Tomoyuki
    • Proceedings of the Korean Society for Emotion and Sensibility Conference
    • /
    • 2000.04a
    • /
    • pp.99-103
    • /
    • 2000
  • There are two approaches in the study of emotion in the physiological psychology. The first is to clarify the brain mechanism of emotion, and the second is to evaluate objectively emotions using physiological responses along with our feeling experience. The method presented here belongs to the second one. Our method is based on the "level-crossing point detection" method. which involves the analysis of frequency fluctuations of EEG and is characterized by estimation of emotionality using coefficients of slopes in the log-power spectra of frequency fluctuation in alpha waves on both the left and right frontal lobe. In this paper we introduce a new theory of estimation on an individual's emotional state by using our non-invasive and easy measurement apparatus.

  • PDF

Emotion Extraction of Multimedia Contents based on Specific Sound Frequency Bands (소리 주파수대역 기반 멀티미디어 콘텐츠의 감성 추출)

  • Kwon, Young-Hun;Chang, Jae-Khun
    • Journal of Digital Convergence
    • /
    • v.11 no.11
    • /
    • pp.381-387
    • /
    • 2013
  • Recently, emotional contents that induce emotions and respond to emotions are given attention in the field of cultural industries, and extracting emotion caused by multimedia contents is being noted. Furthermore, since multimedia contents have been quickly produced and distributed these days, researches automatically to extract the feeling of multimedia contents are being accelerated. In this paper, we will study the method of emotional value extraction in the multimedia contents using the volume value of the multimedia contents in a certain frequency among sound informations. This study allows to extract the emotion of multimedia contents automatically, and the extracted information will be used to provide user's current emotion, weather, etc. for the users.

Research on Designing Korean Emotional Dictionary using Intelligent Natural Language Crawling System in SNS (SNS대상의 지능형 자연어 수집, 처리 시스템 구현을 통한 한국형 감성사전 구축에 관한 연구)

  • Lee, Jong-Hwa
    • The Journal of Information Systems
    • /
    • v.29 no.3
    • /
    • pp.237-251
    • /
    • 2020
  • Purpose The research was studied the hierarchical Hangul emotion index by organizing all the emotions which SNS users are thinking. As a preliminary study by the researcher, the English-based Plutchick (1980)'s emotional standard was reinterpreted in Korean, and a hashtag with implicit meaning on SNS was studied. To build a multidimensional emotion dictionary and classify three-dimensional emotions, an emotion seed was selected for the composition of seven emotion sets, and an emotion word dictionary was constructed by collecting SNS hashtags derived from each emotion seed. We also want to explore the priority of each Hangul emotion index. Design/methodology/approach In the process of transforming the matrix through the vector process of words constituting the sentence, weights were extracted using TF-IDF (Term Frequency Inverse Document Frequency), and the dimension reduction technique of the matrix in the emotion set was NMF (Nonnegative Matrix Factorization) algorithm. The emotional dimension was solved by using the characteristic value of the emotional word. The cosine distance algorithm was used to measure the distance between vectors by measuring the similarity of emotion words in the emotion set. Findings Customer needs analysis is a force to read changes in emotions, and Korean emotion word research is the customer's needs. In addition, the ranking of the emotion words within the emotion set will be a special criterion for reading the depth of the emotion. The sentiment index study of this research believes that by providing companies with effective information for emotional marketing, new business opportunities will be expanded and valued. In addition, if the emotion dictionary is eventually connected to the emotional DNA of the product, it will be possible to define the "emotional DNA", which is a set of emotions that the product should have.

Noise Robust Emotion Recognition Feature : Frequency Range of Meaningful Signal (음성의 특정 주파수 범위를 이용한 잡음환경에서의 감정인식)

  • Kim Eun-Ho;Hyun Kyung-Hak;Kwak Yoon-Keun
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.23 no.5 s.182
    • /
    • pp.68-76
    • /
    • 2006
  • The ability to recognize human emotion is one of the hallmarks of human-robot interaction. Hence this paper describes the realization of emotion recognition. For emotion recognition from voice, we propose a new feature called frequency range of meaningful signal. With this feature, we reached average recognition rate of 76% in speaker-dependent. From the experimental results, we confirm the usefulness of the proposed feature. We also define the noise environment and conduct the noise-environment test. In contrast to other features, the proposed feature is robust in a noise-environment.

The Effect of The Consumers' Emotion Experienced In-Store On Clothing Shopping Behavior According to Shopping Motivation (쇼핑동기에 따른 점포내 소비자 감정이 의류제품 쇼핑행동에 미치는 영향)

  • 정명선;김재숙
    • Journal of the Korean Society of Clothing and Textiles
    • /
    • v.23 no.2
    • /
    • pp.314-325
    • /
    • 1999
  • The purposes of this study were to classify the types of consumers' emotion experienced in-store by shopping motivation and to examined the effects of store environmental factors on emotion and on shopping behavior. The questionnaires were administered to 330 women shopped in department store. Data from 299 women were analyzed by using frequency t-test and regression analysis by SPSS for windows PC program The results of this study were as follows : 1. The consumers' emotion experienced in -store were composed of five factors. But it could be divided by positive negative factors. 2. There was not significant difference in positive emotion between he Product Pu-rchasing Motive Group and the Window Shopping Motive Group. But there was significant difference n negative emotion between two groups. 3. It was found that the effect of environmental factors of apparel store on emotion was significant in both groups. Especially salespeoples' pressure significantly influenced negative emotion in both groups. 4. The emotion experienced in -store significantly influenced clothing shopping behavior.

  • PDF

AM-FM Decomposition and Estimation of Instantaneous Frequency and Instantaneous Amplitude of Speech Signals for Natural Human-robot Interaction (자연스런 인간-로봇 상호작용을 위한 음성 신호의 AM-FM 성분 분해 및 순간 주파수와 순간 진폭의 추정에 관한 연구)

  • Lee, He-Young
    • Speech Sciences
    • /
    • v.12 no.4
    • /
    • pp.53-70
    • /
    • 2005
  • A Vowel of speech signals are multicomponent signals composed of AM-FM components whose instantaneous frequency and instantaneous amplitude are time-varying. The changes of emotion states cause the variation of the instantaneous frequencies and the instantaneous amplitudes of AM-FM components. Therefore, it is important to estimate exactly the instantaneous frequencies and the instantaneous amplitudes of AM-FM components for the extraction of key information representing emotion states and changes in speech signals. In tills paper, firstly a method decomposing speech signals into AM - FM components is addressed. Secondly, the fundamental frequency of vowel sound is estimated by the simple method based on the spectrogram. The estimate of the fundamental frequency is used for decomposing speech signals into AM-FM components. Thirdly, an estimation method is suggested for separation of the instantaneous frequencies and the instantaneous amplitudes of the decomposed AM - FM components, based on Hilbert transform and the demodulation property of the extended Fourier transform. The estimates of the instantaneous frequencies and the instantaneous amplitudes can be used for modification of the spectral distribution and smooth connection of two words in the speech synthesis systems based on a corpus.

  • PDF

Emotion Classification Method Using Various Ocular Features (다양한 눈의 특징 분석을 통한 감성 분류 방법)

  • Kim, Yoonkyoung;Won, Myoung Ju;Lee, Eui Chul
    • The Journal of the Korea Contents Association
    • /
    • v.14 no.10
    • /
    • pp.463-471
    • /
    • 2014
  • In this paper, emotion classification was performed by using four ocular features extracted from near-infrared camera image. According to comparing with previous work, the proposed method used more ocular features and each feature was validated as significant one in terms of emotion classification. To minimize side effects on ocular features caused by using visual stimuli, auditory stimuli for causing two opposite emotion pairs such as "positive-negative" and "arousal-relaxation" were used. As four features for emotion classification, pupil size, pupil accommodation rate, blink frequency, and eye cloased duration were adopted which could be automatically extracted by using lab-made image processing software. At result, pupil accommodation rate and blink frequency were statistically significant features for classification arousal-relaxation. Also, eye closed duration was the most significant feature for classification positive-negative.