• Title/Summary/Keyword: 얼굴 표정 애니메이션

Search Result 91, Processing Time 0.03 seconds

On the Implementation of a Facial Animation Using the Emotional Expression Techniques (FAES : 감성 표현 기법을 이용한 얼굴 애니메이션 구현)

  • Kim Sang-Kil;Min Yong-Sik
    • The Journal of the Korea Contents Association
    • /
    • v.5 no.2
    • /
    • pp.147-155
    • /
    • 2005
  • In this paper, we present a FAES(a Facial Animation with Emotion and Speech) system for speech-driven face animation with emotions. We animate face cartoons not only from input speech, but also based on emotions derived from speech signal. And also our system can ensure smooth transitions and exact representation in animation. To do this, after collecting the training data, we have made the database using SVM(Support Vector Machine) to recognize four different categories of emotions: neutral, dislike, fear and surprise. So that, we can make the system for speech-driven animation with emotions. Also, we trained on Korean young person and focused on only Korean emotional face expressions. Experimental results of our system demonstrate that more emotional areas expanded and the accuracies of the emotional recognition and the continuous speech recognition are respectively increased 7% and 5% more compared with the previous method.

  • PDF

Study of expression in virtual character of facial smile by emotion recognition (감성인식에 따른 가상 캐릭터의 미소 표정변화에 관한 연구)

  • Lee, Dong-Yeop
    • Cartoon and Animation Studies
    • /
    • s.33
    • /
    • pp.383-402
    • /
    • 2013
  • In this study, we apply the facial Facial Action Coding System for coding the muscular system anatomical approach facial expressions to be displayed in response to a change in sensitivity. To verify by applying the virtual character the Duchenne smile to the original. I extracted the Duchenne smile by inducing experiment of emotion (man 2, woman 2) and the movie theater department students trained for the experiment. Based on the expression that has been extracted, I collect the data of the facial muscles. Calculates the frequency of expression of the face and other parts of the body muscles around the mouth and lips, to be applied to the virtual character of the data. Orbicularis muscle to contract end of lips due to shrinkage of the Zygomatic Major is a upward movement, cheek goes up, the movement of the muscles, facial expressions appear the outer eyelid under the eye goes up with a look of smile. Muscle movement of large muscle and surrounding Zygomatic Major is observed together (AU9) muscles around the nose and (AU25, AU26, AU27) muscles around the mouth associated with openness. Duchen smile occurred in the form of Orbicularis Oculi and Zygomatic Major moves at the same time. Based on this, by separating the orbicularis muscle that is displayed in the form of laughter and sympathy to emotional feelings and viable large muscle by the will of the person, by applying to the character of the virtual, and expression of human I try to examine expression of the virtual character's ability to distinguish.

The facial expression generation of vector graphic character using the simplified principle component vector (간소화된 주성분 벡터를 이용한 벡터 그래픽 캐릭터의 얼굴표정 생성)

  • Park, Tae-Hee
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.12 no.9
    • /
    • pp.1547-1553
    • /
    • 2008
  • This paper presents a method that generates various facial expressions of vector graphic character by using the simplified principle component vector. First, we analyze principle components to the nine facial expression(astonished, delighted, etc.) redefined based on Russell's internal emotion state. From this, we find principle component vector having the biggest effect on the character's facial feature and expression and generate the facial expression by using that. Also we create natural intermediate characters and expressions by interpolating weighting values to character's feature and expression. We can save memory space considerably, and create intermediate expressions with a small computation. Hence the performance of character generation system can be considerably improved in web, mobile service and game that real time control is required.

The Facial Expression Controller for 3D Avatar Animation working on a Smartphone (스마트폰기반 3D 아바타 애니메이션을 위한 다양한 얼굴표정 제어기 응용)

  • Choi, In-Ho;Lee, Sang-Hoon;Park, Sang-Il;Kim, Yong-Guk
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2012.06c
    • /
    • pp.323-325
    • /
    • 2012
  • 스마트폰 기반 3D 아바타를 이용하여 임의의 표정을 합성 및 제어하여 애니메이션 할 수 있는 방법과 응용을 제안한다. 사용될 아바타에 표현되는 임의의 표정 Data Set을 PCA로 처리 후, 인간의 가장 기본적인 6 표정으로 컨트롤러 축을 생성한다. 만들어진 제어기에, 임의의 연속 표정을 유저에 의해 정해진 시간에 생성하여 애니메이션 할 수 있는 방법을 시스템을 제안하다. 빠른 계산을 장점으로 하는 본 제어기는 스마트폰 환경에 맞게 탑재 되었으며, 이 제어기를 활용하여 모델워킹 모션에 다양한 표정을 적용할 수 있는 시스템을 구현하였다.

Facial Characteristic Point Extraction for Representation of Facial Expression (얼굴 표정 표현을 위한 얼굴 특징점 추출)

  • Oh, Jeong-Su;Kim, Jin-Tae
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.9 no.1
    • /
    • pp.117-122
    • /
    • 2005
  • This paper proposes an algorithm for Facial Characteristic Point(FCP) extraction. The FCP plays an important role in expression representation for face animation, avatar mimic or facial expression recognition. Conventional algorithms extract the FCP with an expensive motion capture device or by using markers, which give an inconvenience or a psychological load to experimental person. However, the proposed algorithm solves the problems by using only image processing. For the efficient FCP extraction, we analyze and improve the conventional algorithms detecting facial components, which are basis of the FCP extraction.

Study of facial movement around eyes in 3D modeling (3D모델링에서 살펴본 눈 주변의 표정에 대한 연구)

  • 문희정
    • Proceedings of the Korea Multimedia Society Conference
    • /
    • 2003.11b
    • /
    • pp.597-600
    • /
    • 2003
  • 이 연구에서는, 움직임이 커서 확실하게 움직임의 변화를 확인할 수 있는 입 주변의 움직임보다 움직임이 세밀하고 다양해 표현하기가 까다로울 수 있는 눈 주변의 움직임에 있어서 눈 주변의 세밀한 움직임의 근육을 잡는 것과 움직임의 기준정, 또 타겟들의 적절한 조합이 중요한 요소가 된다고 보고 절기에 근거를 두어 블렌드쉐이프를 적용하여 표정 타겟을 만들어 눈 주변의 표정에 대한 연구를 하여 3D얼굴 애니메이션에 적용시키는 방법으로 페이셜 애니메이션 중에서도 눈 주변의 움직임을 중심으로 제안한다.

  • PDF

A Study on Pattern of Facial Expression Presentation in Character Animation (애니메이선 캐릭터의 표정연출 유형 연구)

  • Hong Soon-Koo
    • The Journal of the Korea Contents Association
    • /
    • v.6 no.8
    • /
    • pp.165-174
    • /
    • 2006
  • Birdwhistell explains in the whole communication, language conveys only 35% of the meaning and the rest 65% is conveyed by non-linguistic media. Humans do not entirely depend on linguistic communication, but are sensitive being, using every sense of theirs. Human communication, by using facial expression, gesture as well as language, is able to convey more concrete meaning. Especially, facial expression is a many-sided message system, which delivers Individual Personality, interest, information about response and emotional status, and can be said as powerful communication tool. Though being able to be changed according to various expressive techniques and degree and quality of expression, the symbolic sign of facial expression is characterized by generalized qualify. Animation characters, as roles in story, have vitality by emotional expression of which mental world and psychological status can reveal and read naturally on their actions or facial expressions.

  • PDF

The Multi-marker Tracking for Facial Optical Motion Capture System (Facial Optical Motion Capture System을 위한 다중 마커의 추적)

  • 이문희;김경석
    • Proceedings of the Korea Multimedia Society Conference
    • /
    • 2000.04a
    • /
    • pp.474-477
    • /
    • 2000
  • 최근 3D 애니메이션 , 영화 특수효과 그리고 게임제작시 모션 캡처 시스템(Motion Capture System)을 통하여 실제 인간의 동작 및 표정을 수치적으로 측정해내어 이를 실제 애니메이션에 직접 사용함으로써 막대한 작업시간 및 인력 드리고 자본을 획기적으로 줄이고 있다. 그러나 기존의 모션 캡처 시스템은 고속 카메라를 이용함으로써 가격이 고가이고 움직임 추적에서도 여러 가지 문제점을 가지고 있다. 본 논문에서는 일반 저가의 카메라와 신경회로망 및 영상처리를 이용하여 얼굴 애니메이션용 모션 캡처 시스템에 적용할 수 있는 경제적이고 효율적인 얼굴 움직임 추적 기법을 제안한다.

  • PDF

Data-driven Facial Expression Reconstruction for Simultaneous Motion Capture of Body and Face (동작 및 효정 동시 포착을 위한 데이터 기반 표정 복원에 관한 연구)

  • Park, Sang Il
    • Journal of the Korea Computer Graphics Society
    • /
    • v.18 no.3
    • /
    • pp.9-16
    • /
    • 2012
  • In this paper, we present a new method for reconstructing detailed facial expression from roughly captured data with a small number of markers. Because of the difference in the required capture resolution between the full-body capture and the facial expression capture, they hardly have been performed simultaneously. However, for generating natural animation, a simultaneous capture for body and face is essential. For this purpose, we provide a method for capturing the detailed facial expression only with a small number of markers. Our basic idea is to build a database for the facial expressions and apply the principal component analysis for reducing the dimensionality. The dimensionality reduction enables us to estimate the full data from a part of the data. We justify our method by applying it to dynamic scenes to show the viability of the method.

A Realtime Expression Control for Realistic 3D Facial Animation (현실감 있는 3차원 얼굴 애니메이션을 위한 실시간 표정 제어)

  • Kim Jung-Gi;Min Kyong-Pil;Chun Jun-Chul;Choi Yong-Gil
    • Journal of Internet Computing and Services
    • /
    • v.7 no.2
    • /
    • pp.23-35
    • /
    • 2006
  • This work presents o novel method which extract facial region und features from motion picture automatically and controls the 3D facial expression in real time. To txtract facial region and facial feature points from each color frame of motion pictures a new nonparametric skin color model is proposed rather than using parametric skin color model. Conventionally used parametric skin color models, which presents facial distribution as gaussian-type, have lack of robustness for varying lighting conditions. Thus it needs additional work to extract exact facial region from face images. To resolve the limitation of current skin color model, we exploit the Hue-Tint chrominance components and represent the skin chrominance distribution as a linear function, which can reduce error for detecting facial region. Moreover, the minimal facial feature positions detected by the proposed skin model are adjusted by using edge information of the detected facial region along with the proportions of the face. To produce the realistic facial expression, we adopt Water's linear muscle model and apply the extended version of Water's muscles to variation of the facial features of the 3D face. The experiments show that the proposed approach efficiently detects facial feature points and naturally controls the facial expression of the 3D face model.

  • PDF