• Title/Summary/Keyword: Speech animation

Search Result 42, Processing Time 0.028 seconds

Data-driven Facial Animation Using Sketch Interface (스케치 인터페이스를 이용한 데이터 기반 얼굴 애니메이션)

  • Ju, Eun-Jung;Ahn, Soh-Min;Lee, Je-Hee
    • Journal of the Korea Computer Graphics Society
    • /
    • v.13 no.3
    • /
    • pp.11-18
    • /
    • 2007
  • Creating stylistic facial animation is one of the most important problems in character animation. Traditionally facial animation is created manually by animators of captured using motion capture systems. But this process is very difficult and labor-intensive. In this work, we present an intuitive, easy-to-use, sketch-based user interface system that facilitates the process of creating facial animation and key-frame interpolation method using facial capture data. The user of our system is allowed to create expressive speech facial animation easily and rapidly.

  • PDF

MPEG-4TTS 현황 및 전망

  • 한민수
    • The Magazine of the IEIE
    • /
    • v.24 no.9
    • /
    • pp.91-98
    • /
    • 1997
  • Text-to-Speech(WS) technology has been attracting a lot of interest among speech engineers because of its own benefits. Namely, the possible application areas of talking computers, emergency alarming systems in speech, speech output devices for speech-impaired, and so on. Hence, many researchers have made significant progresses in the speech synthesis techniques in the sense of their own languages and as a result, the quality of current speech synthesizers are believed to be acceptable to normal users. These are partly why the MPEG group had decided to include the WS technology as one of its MPEG-4 functionalities. ETRI has made major contributions to the current MPEG-4 775 appearing in various MPEG-4 documents with relatively minor contributions from AT&T and NW. Main MPEG-4 functionalities presently available are; 1) use of original prosody for synthesized speech output, 2) trick mode functions for general users without breaking synthesized speech prosody, 3) interoperability with Facial Animation(FA) tools, and 4) dubbing a moving/anlmated picture with lip-shape pattern informations.

  • PDF

Speech Animation with Multilevel Control (다중 제어 레벨을 갖는 입모양 중심의 표정 생성)

  • Moon, Bo-Hee;Lee, Son-Ou;Wohn, Kwang-yun
    • Korean Journal of Cognitive Science
    • /
    • v.6 no.2
    • /
    • pp.47-79
    • /
    • 1995
  • Since the early age of computer graphics, facial animation has been applied to various fields, and nowadays it has found several novel applications such as virtual reality(for representing virtual agents), teleconference, and man-machine interface.When we want to apply facial animation to the system with multiple participants connected via network, it is hard to animate facial expression as we desire in real-time because of the size of information to maintain an efficient communication.This paper's major contribution is to adapt 'Level-of-Detail'to the facial animation in order to solve the above problem.Level-of-Detail has been studied in the field of computer graphics to reperesent the appearance of complicated objects in efficient and adaptive way, but until now no attempt has mode in the field of facial animation. In this paper, we present a systematic scheme which enables this kind of adaptive control using Level-of-Detail.The implemented system can generate speech synchronized facial expressions with various types of user input such as text, voice, GUI, head motion, etc.

  • PDF

HEEAS: On the Implementation and an Animation Algorithm of an Emotional Expression (HEEAS: 감정표현 애니메이션 알고리즘과 구현에 관한 연구)

  • Kim Sang-Kil;Min Yong-Sik
    • The Journal of the Korea Contents Association
    • /
    • v.6 no.3
    • /
    • pp.125-134
    • /
    • 2006
  • The purpose of this paper is constructed a HEEAAS(Human Emotional Expression Animaion System), which is an animation system to show both the face and the body motion from the inputted voice about just 4 types of emotions such as fear, dislike, surprise and normal. To implement our paper, we chose the korean young man in his twenties who was to show appropriate emotions the most correctly. Also, we have focused on reducing the processing time about making the real animation in making both face and body codes of emotions from the inputted voice signal. That is, we can reduce the search time to use the binary search technique from the face and body motion databases, Throughout the experiment, we have a 99.9% accuracy of the real emotional expression in the cartoon animation.

  • PDF

Coarticulation Model of Hangul Visual speedh for Lip Animation (입술 애니메이션을 위한 한글 발음의 동시조음 모델)

  • Gong, Gwang-Sik;Kim, Chang-Heon
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.26 no.9
    • /
    • pp.1031-1041
    • /
    • 1999
  • 기존의 한글에 대한 입술 애니메이션 방법은 음소의 입모양을 몇 개의 입모양으로 정의하고 이들을 보간하여 입술을 애니메이션하였다. 하지만 발음하는 동안의 실제 입술 움직임은 선형함수나 단순한 비선형함수가 아니기 때문에 보간방법에 의해 중간 움직임을 생성하는 방법으로는 음소의 입술 움직임을 효과적으로 생성할 수 없다. 또 이 방법은 동시조음도 고려하지 않아 음소들간에 변화하는 입술 움직임도 표현할 수 없었다. 본 논문에서는 동시조음을 고려하여 한글을 자연스럽게 발음하는 입술 애니메이션 방법을 제안한다. 비디오 카메라로 발음하는 동안의 음소의 움직임들을 측정하고 입술 움직임 제어 파라미터들을 추출한다. 각각의 제어 파라미터들은 L fqvist의 스피치 생성 제스처 이론(speech production gesture theory)을 이용하여 실제 음소의 입술 움직임에 근사한 움직임인 지배함수(dominance function)들로 정의되고 입술 움직임을 애니메이션할 때 사용된다. 또, 각 지배함수들은 혼합함수(blending function)와 반음절에 의한 한글 합성 규칙을 사용하여 결합하고 동시조음이 적용된 한글을 발음하게 된다. 따라서 스피치 생성 제스처 이론을 이용하여 입술 움직임 모델을 구현한 방법은 기존의 보간에 의해 중간 움직임을 생성한 방법보다 실제 움직임에 근사한 움직임을 생성하고 동시조음도 고려한 움직임을 보여준다.Abstract The existing lip animation method of Hangul classifies the shape of lips with a few shapes and implements the lip animation with interpolating them. However it doesn't represent natural lip animation because the function of the real motion of lips, during articulation, isn't linear or simple non-linear function. It doesn't also represent the motion of lips varying among phonemes because it doesn't consider coarticulation. In this paper we present a new coarticulation model for the natural lip animation of Hangul. Using two video cameras, we film the speaker's lips and extract the lip control parameters. Each lip control parameter is defined as dominance function by using L fqvist's speech production gesture theory. This dominance function approximates to the real lip animation of a phoneme during articulation of one and is used when lip animation is implemented. Each dominance function combines into blending function by using Hangul composition rule based on demi-syllable. Then the lip animation of our coarticulation model represents natural motion of lips. Therefore our coarticulation model approximates to real lip motion rather than the existing model and represents the natural lip motion considered coarticulation.

The Analysis for the Distinctive Directing of Speech Balloons in Webtoon (웹툰에 나타난 특징적 말칸 연출에 대한 분석)

  • Jeung, Kiu-Ha;Yoon, Ki-Heon
    • Cartoon and Animation Studies
    • /
    • s.36
    • /
    • pp.393-416
    • /
    • 2014
  • Comics has three components: cuts, gap between cuts, speech balloons. Still, it is true that speech balloons are not commonly subject to the study for comics. A few preceding researches pinpoint exactly the morphological features and functions of speech balloons. In today when webtoon becomes generalized, these features and functions are continued as they are and are used in webtoon. We can catch that speech balloons are also affected since the environmental elements of web induce the change in the overall comics directing. There are two perspectives to sort out the features of speech balloons: first, the placement issue of speech balloons. The unlimited expansion of web space gives the environment for comicss to use the gap between cuts as wide as they can. It leads to turn out some of the ways to place the balloons, so we can sort them out general placement, exterior placement, the upper and lower placement, scroll-use type. Second, as the directing techniques for webtoon become digitalized by the morphological issue, speech balloon itself has been expanded its ways to express by various expression methods. Analyzing and classifying, recording the newly emerged conditions on the preceding study are worthy of trying and will become the cornerstone for the follow research.

The Study on Body Language in Animation as Functional Aspects -Focusing on Mulan, Beauty and the beast, Aladdin, Sinbad- (기능론적 관점에서 본 애니메이션의 신체언어 연구 - 뮬란, 미녀와 야수, 알라딘, 신밧드를 중심으로-)

  • Chung, Mi-Ghang;Lee, Mi-Young;Kim, Sung-Hee;Kim, Jae-Ho
    • Archives of design research
    • /
    • v.20 no.1 s.69
    • /
    • pp.55-64
    • /
    • 2007
  • Non-verbal communications are important because they support and replace verbal communication. Body language of various non-verbal communications is the communication using the body. In animation, expression of body language is very important because characters play an important role in communicating the scenario. Animation has a dual communication structure, different from general communication. One is the communication between the speaker character and the hearer character, the other is the image and the audience, which includes the communication between the speaker character and the hearer character. In this study, we divide the body language from the characters into the discourse-in act and discourse-out act according to this dual structure and classify it into adaptors, emblem, illustrator, regulator, affect display by a functional approach method. Especially, the illustrator is subdivided into pragmatic speech act. Finally, this study analyzes the features of body language in animation and represents animation character's body language for an effective expression of the communications in animation.

  • PDF

Development of Automatic Lip-sync MAYA Plug-in for 3D Characters (3D 캐릭터에서의 자동 립싱크 MAYA 플러그인 개발)

  • Lee, Sang-Woo;Shin, Sung-Wook;Chung, Sung-Taek
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.18 no.3
    • /
    • pp.127-134
    • /
    • 2018
  • In this paper, we have developed the Auto Lip-Sync Maya plug-in for extracting Korean phonemes from voice data and text information based on Korean and produce high quality 3D lip-sync animation using divided phonemes. In the developed system, phoneme separation was classified into 8 vowels and 13 consonants used in Korean, referring to 49 phonemes provided by Microsoft Speech API engine SAPI. In addition, the pronunciation of vowels and consonants has variety Mouth Shapes, but the same Viseme can be applied to some identical ones. Based on this, we have developed Auto Lip-sync Maya Plug-in based on Python to enable lip-sync animation to be implemented automatically at once.

A Study on Aesthetic Characteristics and social communication of Korean Independent Animation (한국 독립애니메이션의 미학적 특성과 사회적 소통방식 연구 - '인디애니페스트' 수상작 중심으로 -)

  • Seo, Soo-jung
    • Cartoon and Animation Studies
    • /
    • s.47
    • /
    • pp.125-148
    • /
    • 2017
  • This paper analyzes Korean aesthetic characteristics of Korean independent animation, which is closely related to artistic and social characteristics, by searching Korean independent animation image for the winner of "Indie AniFest", the only independent animation film festival in Korea. And social communication methods. There are three ways in which Korean Independence Animation shows the way of speaking through award-winning works. The first is a method of expressing feelings, impulses, and desires in a formative way, and a direct and sensuous image as in sign language or sound. In the second method of speaking differently with images, in the third method, It makes the voice of the fringe through the real world of the story sound as a story of reality. Animation is a medium that 'originalizes and communicates meaning' as an image, and has implemented and constructed a new way of speaking, which is different from existing social voices. Nevertheless, it is not easy to find a research that approaches animation from the viewpoint of social and political speech through images. Therefore, this paper is necessary for the balanced development of animation. In addition, this study can contribute to re - examining Korean independent animation from an academic point of view and to discover and evaluate fair value from a wider perspective.

Speaker Adapted Real-time Dialogue Speech Recognition Considering Korean Vocal Sound System (한국어 음운체계를 고려한 화자적응 실시간 단모음인식에 관한 연구)

  • Hwang, Seon-Min;Yun, Han-Kyung;Song, Bok-Hee
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.6 no.4
    • /
    • pp.201-207
    • /
    • 2013
  • Voice Recognition technique has been developed and it has been actively applied to various information devices such as smart phones and car navigation system. But the basic research technique related the speech recognition is based on research results in English. Since the lip sync producing generally requires tedious hand work of animators and it serious affects the animation producing cost and development period to get a high quality lip animation. In this research, a real time processed automatic lip sync algorithm for virtual characters in digital contents is studied by considering Korean vocal sound system. This suggested algorithm contributes to produce a natural lip animation with the lower producing cost and the shorter development period.