• Title/Summary/Keyword: 음성 감성 인식

Search Result 52, Processing Time 0.031 seconds

Adaptive Noise Reduction using Standard Deviation of Wavelet Coefficients in Speech Signal (웨이브렛 계수의 표준편차를 이용한 음성신호의 적응 잡음 제거)

  • 황향자;정광일;이상태;김종교
    • Science of Emotion and Sensibility
    • /
    • v.7 no.2
    • /
    • pp.141-148
    • /
    • 2004
  • This paper proposed a new time adapted threshold using the standard deviations of Wavelet coefficients after Wavelet transform by frame scale. The time adapted threshold is set up using the sum of standard deviations of Wavelet coefficient in cA3 and weighted cDl. cA3 coefficients represent the voiced sound with low frequency and cDl coefficients represent the unvoiced sound with high frequency. From simulation results, it is demonstrated that the proposed algorithm improves SNR and MSE performance more than Wavelet transform and Wavelet packet transform does. Moreover, the reconstructed signals by the proposed algorithm resemble the original signal in terms of plosive sound, fricative sound and affricate sound but Wavelet transform and Wavelet packet transform reduce those sounds seriously.

  • PDF

A Study on Algorithm of Emotion Analysis using EEG and HRV (뇌전도와 심박변이를 이용한 감성 분석 알고리즘에 대한 연구)

  • Chon, Ki-Hwan;Oh, Ju-Young;Park, Sun-Hee;Jeong, Yeon-Man;Yang, Dong-Il
    • Journal of the Korea Society of Computer and Information
    • /
    • v.15 no.10
    • /
    • pp.105-112
    • /
    • 2010
  • In this paper, the bio-signals, such as EEG, ECG were measured with a sensor and their characters were drawn out and analyzed. With results from the analysis, four emotion of rest, concentration, tension and depression were inferred. In order to assess one's emotion, the characteristic vectors were drawn out by applying various ways, including the frequency analysis of the bio-signals like the measured EEG and HRV. RBFN, a neural network of the complex structure of unsupervised and supervised learning, was applied to classify and infer the deducted information. Through experiments, the system suggested in this thesis showed better capability to classify and infer than other systems using a different neural network. As follow-up research tasks, the recognizance rate of the measured bio-signals should be improved. Also, the technology which can be applied to the wired or wireless sensor measuring the bio-signals more easily and to wearable computing should be developed.

A study on the enhancement of emotion recognition through facial expression detection in user's tendency (사용자의 성향 기반의 얼굴 표정을 통한 감정 인식률 향상을 위한 연구)

  • Lee, Jong-Sik;Shin, Dong-Hee
    • Science of Emotion and Sensibility
    • /
    • v.17 no.1
    • /
    • pp.53-62
    • /
    • 2014
  • Despite the huge potential of the practical application of emotion recognition technologies, the enhancement of the technologies still remains a challenge mainly due to the difficulty of recognizing emotion. Although not perfect, human emotions can be recognized through human images and sounds. Emotion recognition technologies have been researched by extensive studies that include image-based recognition studies, sound-based studies, and both image and sound-based studies. Studies on emotion recognition through facial expression detection are especially effective as emotions are primarily expressed in human face. However, differences in user environment and their familiarity with the technologies may cause significant disparities and errors. In order to enhance the accuracy of real-time emotion recognition, it is crucial to note a mechanism of understanding and analyzing users' personality traits that contribute to the improvement of emotion recognition. This study focuses on analyzing users' personality traits and its application in the emotion recognition system to reduce errors in emotion recognition through facial expression detection and improve the accuracy of the results. In particular, the study offers a practical solution to users with subtle facial expressions or low degree of emotion expression by providing an enhanced emotion recognition function.

A Study on the Ubiquitous Home Network Interface System by Application of User's Gesture Recognition Method (사용자 제스처 인식을 활용한 유비쿼터스 홈 네트워크 인터페이스 체계에 대한 연구)

  • Park In-Chan;Kim Sun-Chul
    • Science of Emotion and Sensibility
    • /
    • v.8 no.3
    • /
    • pp.265-276
    • /
    • 2005
  • 현재의 유비쿼터스 환경의 홈 네트워크 제품 사용자는 단일 사용자가 아닌 다수의 사용자가 사용하는 네트워크 행태를 취하고 있다. 변화하는 사용환경과 시스템들은 현재와는 다른 요구사항을 가지고 있으며, 이에 따른 사용자 중심의 디자인과 제품 인터페이스 체계의 연구활동은 국내외에서 활발하게 이루어지고 있다. 다양한 모바일 디바이스 및 홈 네트워크 제품의 보급화가 빠르게 성장하면서 이를 쉽게 제어하기 위한 다양한 제어방식이 연구되고 있다. 이중 음성인식기술을 비롯한 표정은 안면표정인식기술의 개발이 활발히 진행되고 있다. 모션감지 센서를 활용한 사용자 제스처 콘트롤 체계는 아직까지는 초보적인 단계에 있으나, 제품 제어에 있어서 향후 근미래에는 자연스러운 인터랙티브 인터페이스의 활용도가 높아질 전망이다. 이에 본 연구에서는 효과적인 디바이스 제어를 위한 제스처 유형의 자연스러운 사용언어체계 개발 방법 및 결과 그리고 사용자 맨탈모델와 메타포 실험을 통한 연구내용을 정리하였다. 기존 사용자의 제스처 유형의 자연스러운 사용언어를 분석하면서 디바이스 제어방식으로서 활용 가능성을 검토할 수 있었으며, 동작 감지 카메라 및 센서를 활용한 새로운 디바이스 제어방식 개발과정의 연구를 통하여 제스처 유형의 자연스러운 언어 체계 개발 및 과정을 정립하였다.

  • PDF

Design for Mood-Matched Music Based on Deep Learning Emotion Recognition (딥러닝 감정 인식 기반 배경음악 매칭 설계)

  • Chung, Moonsik;Moon, Nammee
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2021.11a
    • /
    • pp.834-836
    • /
    • 2021
  • 멀티모달 감정인식을 통해 사람의 감정을 정확하게 분류하고, 사람의 감정에 어울리는 음악을 매칭하는 시스템을 설계한다. 멀티모달 감정 인식 방법으로는 IEMOCAP(Interactive Emotional Dyadic Motion Capture) 데이터셋을 활용해 감정을 분류하고, 분류된 감정의 분위기에 맞는 음악을 매칭시키는 시스템을 구축하고자 한다. 유니모달 대비 멀티모달 감정인식의 정확도를 개선한 시스템을 통해 텍스트, 음성, 표정을 포함하고 있는 동영상의 감성 분위기에 적합한 음악 매칭 시스템을 연구한다.

The Emotion Recognition System through The Extraction of Emotional Components from Speech (음성의 감성요소 추출을 통한 감성 인식 시스템)

  • Park Chang-Hyun;Sim Kwee-Bo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.10 no.9
    • /
    • pp.763-770
    • /
    • 2004
  • The important issue of emotion recognition from speech is a feature extracting and pattern classification. Features should involve essential information for classifying the emotions. Feature selection is needed to decompose the components of speech and analyze the relation between features and emotions. Specially, a pitch of speech components includes much information for emotion. Accordingly, this paper searches the relation of emotion to features such as the sound loudness, pitch, etc. and classifies the emotions by using the statistic of the collecting data. This paper deals with the method of recognizing emotion from the sound. The most important emotional component of sound is a tone. Also, the inference ability of a brain takes part in the emotion recognition. This paper finds empirically the emotional components from the speech and experiment on the emotion recognition. This paper also proposes the recognition method using these emotional components and the transition probability.

Recognition of Emotional State of Speaker Using Machine learning (SVM 을 이용한 화자의 감정상태 인식)

  • Lee, Na-Ra;Choi, Hoon-Ha;Kim, Hyun-jung;Won, Il-Young
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2012.11a
    • /
    • pp.468-471
    • /
    • 2012
  • 음성을 통한 자동화된 감정 인식은 편리하고 다양한 서비스를 제공할 수 있어 중요한 연구분야라고 할 수 있다. 기계학습의 다양한 알고리즘을 사용하여 감정을 인식하는 연구가 진행되어 왔지만 그 성능은 아직 초보적 단계를 벋어나지 못하고 있는 실정이다. 앞선 연구에서 우리는 비감독 학습 방법으로 감성을 그룹화 하고 이것을 이용하여 다시 감독 학습을 하는 시스템을 소개 하였다. 본 연구에서 우리는 감독 학습 방법에서 사용했던 오류 역전파 알고리즘을 support vector machine(SVM) 으로 변경하고 몇 가지 구조를 변경하여 기능을 개선하였다. 실험을 통하여 성능을 측정하였으며 어느 정도 개선된 결과를 얻을 수 있었다.

HEEAS: On the Implementation and an Animation Algorithm of an Emotional Expression (HEEAS: 감정표현 애니메이션 알고리즘과 구현에 관한 연구)

  • Kim Sang-Kil;Min Yong-Sik
    • The Journal of the Korea Contents Association
    • /
    • v.6 no.3
    • /
    • pp.125-134
    • /
    • 2006
  • The purpose of this paper is constructed a HEEAAS(Human Emotional Expression Animaion System), which is an animation system to show both the face and the body motion from the inputted voice about just 4 types of emotions such as fear, dislike, surprise and normal. To implement our paper, we chose the korean young man in his twenties who was to show appropriate emotions the most correctly. Also, we have focused on reducing the processing time about making the real animation in making both face and body codes of emotions from the inputted voice signal. That is, we can reduce the search time to use the binary search technique from the face and body motion databases, Throughout the experiment, we have a 99.9% accuracy of the real emotional expression in the cartoon animation.

  • PDF

The Pattern Recognition Methods for Emotion Recognition with Speech Signal (음성신호를 이용한 감성인식에서의 패턴인식 방법)

  • Park Chang-Hyun;Sim Kwee-Bo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.12 no.3
    • /
    • pp.284-288
    • /
    • 2006
  • In this paper, we apply several pattern recognition algorithms to emotion recognition system with speech signal and compare the results. Firstly, we need emotional speech databases. Also, speech features for emotion recognition is determined on the database analysis step. Secondly, recognition algorithms are applied to these speech features. The algorithms we try are artificial neural network, Bayesian learning, Principal Component Analysis, LBG algorithm. Thereafter, the performance gap of these methods is presented on the experiment result section. Truly, emotion recognition technique is not mature. That is, the emotion feature selection, relevant classification method selection, all these problems are disputable. So, we wish this paper to be a reference for the disputes.

The Pattern Recognition Methods for Emotion Recognition with Speech Signal (음성신호를 이용한 감성인식에서의 패턴인식 방법)

  • Park Chang-Hyeon;Sim Gwi-Bo
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2006.05a
    • /
    • pp.347-350
    • /
    • 2006
  • In this paper, we apply several pattern recognition algorithms to emotion recognition system with speech signal and compare the results. Firstly, we need emotional speech databases. Also, speech features for emotion recognition is determined on the database analysis step. Secondly, recognition algorithms are applied to these speech features. The algorithms we try are artificial neural network, Bayesian learning, Principal Component Analysis, LBG algorithm. Thereafter, the performance gap of these methods is presented on the experiment result section. Truly, emotion recognition technique is not mature. That is, the emotion feature selection, relevant classification method selection, all these problems are disputable. So, we wish this paper to be a reference for the disputes.

  • PDF