• Title/Summary/Keyword: 얼굴표정

Search Result 518, Processing Time 0.026 seconds

Study of expression in virtual character of facial smile by emotion recognition (감성인식에 따른 가상 캐릭터의 미소 표정변화에 관한 연구)

  • Lee, Dong-Yeop
    • Cartoon and Animation Studies
    • /
    • s.33
    • /
    • pp.383-402
    • /
    • 2013
  • In this study, we apply the facial Facial Action Coding System for coding the muscular system anatomical approach facial expressions to be displayed in response to a change in sensitivity. To verify by applying the virtual character the Duchenne smile to the original. I extracted the Duchenne smile by inducing experiment of emotion (man 2, woman 2) and the movie theater department students trained for the experiment. Based on the expression that has been extracted, I collect the data of the facial muscles. Calculates the frequency of expression of the face and other parts of the body muscles around the mouth and lips, to be applied to the virtual character of the data. Orbicularis muscle to contract end of lips due to shrinkage of the Zygomatic Major is a upward movement, cheek goes up, the movement of the muscles, facial expressions appear the outer eyelid under the eye goes up with a look of smile. Muscle movement of large muscle and surrounding Zygomatic Major is observed together (AU9) muscles around the nose and (AU25, AU26, AU27) muscles around the mouth associated with openness. Duchen smile occurred in the form of Orbicularis Oculi and Zygomatic Major moves at the same time. Based on this, by separating the orbicularis muscle that is displayed in the form of laughter and sympathy to emotional feelings and viable large muscle by the will of the person, by applying to the character of the virtual, and expression of human I try to examine expression of the virtual character's ability to distinguish.

Effects of the facial expression presenting types and facial areas on the emotional recognition (얼굴 표정의 제시 유형과 제시 영역에 따른 정서 인식 효과)

  • Lee, Jung-Hun;Park, Soo-Jin;Han, Kwang-Hee;Ghim, Hei-Rhee;Cho, Kyung-Ja
    • Science of Emotion and Sensibility
    • /
    • v.10 no.1
    • /
    • pp.113-125
    • /
    • 2007
  • The aim of the experimental studies described in this paper is to investigate the effects of the face/eye/mouth areas using dynamic facial expressions and static facial expressions on emotional recognition. Using seven-seconds-displays, experiment 1 for basic emotions and experiment 2 for complex emotions are executed. The results of two experiments supported that the effects of dynamic facial expressions are higher than static one on emotional recognition and indicated the higher emotional recognition effects of eye area on dynamic images than mouth area. These results suggest that dynamic properties should be considered in emotional study with facial expressions for not only basic emotions but also complex emotions. However, we should consider the properties of emotion because each emotion did not show the effects of dynamic image equally. Furthermore, this study let us know which facial area shows emotional states more correctly is according to the feature emotion.

  • PDF

Reconstruction from Feature Points of Face through Fuzzy C-Means Clustering Algorithm with Gabor Wavelets (FCM 군집화 알고리즘에 의한 얼굴의 특징점에서 Gabor 웨이브렛을 이용한 복원)

  • 신영숙;이수용;이일병;정찬섭
    • Korean Journal of Cognitive Science
    • /
    • v.11 no.2
    • /
    • pp.53-58
    • /
    • 2000
  • This paper reconstructs local region of a facial expression image from extracted feature points of facial expression image using FCM(Fuzzy C-Meang) clustering algorithm with Gabor wavelets. The feature extraction in a face is two steps. In the first step, we accomplish the edge extraction of main components of face using average value of 2-D Gabor wavelets coefficient histogram of image and in the next step, extract final feature points from the extracted edge information using FCM clustering algorithm. This study presents that the principal components of facial expression images can be reconstructed with only a few feature points extracted from FCM clustering algorithm. It can also be applied to objects recognition as well as facial expressions recognition.

  • PDF

A Face Recognition Method Robust to Variations in Lighting and Facial Expression (조명 변화, 얼굴 표정 변화에 강인한 얼굴 인식 방법)

  • Yang, Hui-Seong;Kim, Yu-Ho;Lee, Jun-Ho
    • Journal of KIISE:Software and Applications
    • /
    • v.28 no.2
    • /
    • pp.192-200
    • /
    • 2001
  • 본 논문은 조명 변화, 표정 변화, 부분적인 오클루전이 있는 얼굴 영상에 강인하고 적은 메모리양과 계산량을 갖는 효율적인 얼굴 인식 방법을 제안한다. SKKUface(Sungkyunkwan University face)라 명명한 이 방법은 먼저 훈련 영상에 PCA(principal component analysis)를 적용하여 차원을 줄일 때 구해지는 특징 벡터 공간에서 조명 변화, 얼굴 표정 변화 등에 해당되는 공간이 최대한 제외된 새로운 특징 벡터 공간을 생성한다. 이러한 특징 벡터 공간은 얼굴의 고유특징만을 주로 포함하는 벡터 공간이므로 이러한 벡터 공간에 Fisher linear discriminant를 적용하면 클래스간의 더욱 효과적인 분리가 이루어져 인식률을 획기적으로 향상시킨다. 또한, SKKUface 방법은 클래스간 분산(between-class covariance) 행렬과 클래스내 분산(within-class covariance) 행렬을 계산할 때 문제가 되는 메모리양과 계산 시간을 획기적으로 줄이는 방법을 제안하여 적용하였다. 제안된 SKKUface 방법의 얼굴 인식 성능을 평가하기 위하여 YALE, SKKU, ORL(Olivetti Research Laboratory) 얼굴 데이타베이스를 가지고 기존의 얼굴 인식 방법으로 널리 알려진 Eigenface 방법, Fisherface 방법과 함께 인식률을 비교 평가하였다. 실험 결과, 제안된 SKKUface 방법이 조명 변화, 부분적인 오클루전이 있는 얼굴 영상에 대해서 Eigenface 방법과 Fisherface 방법에 비해 인식률이 상당히 우수함을 알 수 있었다.

  • PDF

A Study on Appearance-Based Facial Expression Recognition Using Active Shape Model (Active Shape Model을 이용한 외형기반 얼굴표정인식에 관한 연구)

  • Kim, Dong-Ju;Shin, Jeong-Hoon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.1
    • /
    • pp.43-50
    • /
    • 2016
  • This paper introduces an appearance-based facial expression recognition method using ASM landmarks which is used to acquire a detailed face region. In particular, EHMM-based algorithm and SVM classifier with histogram feature are employed to appearance-based facial expression recognition, and performance evaluation of proposed method was performed with CK and JAFFE facial expression database. In addition, performance comparison was achieved through comparison with distance-based face normalization method and a geometric feature-based facial expression approach which employed geometrical features of ASM landmarks and SVM algorithm. As a result, the proposed method using ASM-based face normalization showed performance improvements of 6.39% and 7.98% compared to previous distance-based face normalization method for CK database and JAFFE database, respectively. Also, the proposed method showed higher performance compared to geometric feature-based facial expression approach, and we confirmed an effectiveness of proposed method.

A Study on Face Detection Performance Enhancement Using FLD (FLD를 이용한 얼굴 검출의 성능 향상을 위한 연구)

  • 남미영;이필규;김광백
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2004.04a
    • /
    • pp.225-230
    • /
    • 2004
  • 얼굴 검출은 디지털화된 임의의 정지 영상 혹은 연속된 영상으로부터 얼굴 존재 유무를 판단하고, 얼굴이 존재할 경우 영상 내 얼굴의 위치, 방향, 크기 둥을 알아내는 기술로 정의된다. 이러한 얼굴 검출은 얼굴 인식이나 표정인식, 헤드 재스쳐 등의 기초 기술로서 해당 시스템의 성능에 매우 중요한 변수 중에 하나이다. 그러나 영상내의 얼굴은 표정, 포즈, 크기, 빛의 방향 및 밝기, 안경, 수염 둥의 환경적 변화로 인해 얼굴 모양이 다양해지므로 정확하고 빠른 검출이 어렵다. 따라서 본 논문에서는 피셔의 선형 판별 분석을 이용하여 몇 가지 환경적 조건을 극복한 정확하고 빠른 얼굴 검출 방법을 제안한다. 제안된 방법은 포즈와, 배경에 무관하게 얼굴을 검출하면서도 빠른 검출이 가능하다. 이를 위해 계층적인 방법으로 얼굴 검출을 수행하며, 휴리스틱한 방법, 피셔의 판별 분석을 이용하여 얼굴 검출을 수행하고 검색 영역의 축소와 선형 결정의 계산 시간의 단축으로 검출 응답 시간을 빠르게 하였다 추출된 얼굴 영상에서 포즈를 추정하고 눈 영역을 검출함으로써 얼굴 정보의 사용에 있어 보다 많은 정보를 추출할 수 있도록 하였다.

  • PDF

Synthesis of Realistic Facial Expression using a Nonlinear Model for Skin Color Change (비선형 피부색 변화 모델을 이용한 실감적인 표정 합성)

  • Lee Jeong-Ho;Park Hyun;Moon Young-Shik
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.43 no.3 s.309
    • /
    • pp.67-75
    • /
    • 2006
  • Facial expressions exhibit not only facial feature motions, but also subtle changes in illumination and appearance. Since it is difficult to generate realistic facial expressions by using only geometric deformations, detailed features such as textures should also be deformed to achieve more realistic expression. The existing methods such as the expression ratio image have drawbacks, in that detailed changes of complexion by lighting can not be generated properly. In this paper, we propose a nonlinear model for skin color change and a model-based synthesis method for facial expression that can apply realistic expression details under different lighting conditions. The proposed method is composed of the following three steps; automatic extraction of facial features using active appearance model and geometric deformation of expression using warping, generation of facial expression using a model for nonlinear skin color change, and synthesis of original face with generated expression using a blending ratio that is computed by the Euclidean distance transform. Experimental results show that the proposed method generate realistic facial expressions under various lighting conditions.

Hierarchical Visualization of the Space of Facial Expressions (얼굴 표정공간의 계층적 가시화)

  • Kim Sung-Ho;Jung Moon-Ryul
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.31 no.12
    • /
    • pp.726-734
    • /
    • 2004
  • This paper presents a facial animation method that enables the user to select a sequence of facial frames from the facial expression space, whose level of details the user can select hierarchically Our system creates the facial expression space from about 2400 captured facial frames. To represent the state of each expression, we use the distance matrix that represents the distance between pairs of feature points on the face. The shortest trajectories are found by dynamic programming. The space of facial expressions is multidimensional. To navigate this space, we visualize the space of expressions in 2D space by using the multidimensional scaling(MDS). But because there are too many facial expressions to select from, the user faces difficulty in navigating the space. So, we visualize the space hierarchically. To partition the space into a hierarchy of subspaces, we use fuzzy clustering. In the beginning, the system creates about 10 clusters from the space of 2400 facial expressions. Every tine the level increases, the system doubles the number of clusters. The cluster centers are displayed on 2D screen and are used as candidate key frames for key frame animation. The user selects new key frames along the navigation path of the previous level. At the maximum level, the user completes key frame specification. We let animators use the system to create example animations, and evaluate the system based on the results.

Effect of Depressive Mood on Identification of Emotional Facial Expression (우울감이 얼굴 표정 정서 인식에 미치는 영향)

  • Ryu, Kyoung-Hi;Oh, Kyung-Ja
    • Science of Emotion and Sensibility
    • /
    • v.11 no.1
    • /
    • pp.11-21
    • /
    • 2008
  • This study was designed to examine the effect of depressive mood on identification of emotional facial expression. Participants were screened out of 305 college students on the basis of the BDI-II score. Students with BDI-II score higher than 14(upper 20%) were selected for the Depression Group and those with BDI-II score lower than 5(lower 20%) were selected for the Control Group. A final sample of 20 students in the Depression Group and 20 in the Control Group were presented with facial expression stimuli of an increasing degree of emotional intensity, slowly changing from a neutral to a full intensity of happy, sad, angry, or fearful expressions. The result showed that there was the significant interaction of Group by Emotion(esp. happy and sad) which suggested that depressive mood affects processing of emotional stimuli such as facial expressions. Implication of this result for mood-congruent information processing were discussed.

  • PDF

Facial Expression Recognition Using SIFT Descriptor (SIFT 기술자를 이용한 얼굴 표정인식)

  • Kim, Dong-Ju;Lee, Sang-Heon;Sohn, Myoung-Kyu
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.2
    • /
    • pp.89-94
    • /
    • 2016
  • This paper proposed a facial expression recognition approach using SIFT feature and SVM classifier. The SIFT was generally employed as feature descriptor at key-points in object recognition fields. However, this paper applied the SIFT descriptor as feature vector for facial expression recognition. In this paper, the facial feature was extracted by applying SIFT descriptor at each sub-block image without key-point detection procedure, and the facial expression recognition was performed using SVM classifier. The performance evaluation was carried out through comparison with binary pattern feature-based approaches such as LBP and LDP, and the CK facial expression database and the JAFFE facial expression database were used in the experiments. From the experimental results, the proposed method using SIFT descriptor showed performance improvements of 6.06% and 3.87% compared to previous approaches for CK database and JAFFE database, respectively.