• Title/Summary/Keyword: 입술 추출

Search Result 76, Processing Time 0.027 seconds

A Design and Implementation of Gesture Recognition System (제스쳐 인식 시스템 설계 및 구현)

  • Kim, Kun-Woo;Kim, Chang-Hyun;Jeon, Chang-Ho;Lee, Won-Joo
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2008.06a
    • /
    • pp.231-235
    • /
    • 2008
  • 컴퓨터 및 주변기기의 성능이 발전함에 따라 영상처리에 대한 관심이 높아지고, 영상으로부터 원하는 정보를 얻기 위한 연구가 활발히 진행되고 있다. 이러한 연구에서 움직임 추적, 특정 사물 추출, 동영상 검색 등으로 정보를 추출하는 과정은 높은 시스템 자원을 요구하기 때문에 멀티태스킹이 어렵다. 따라서 본 논문에서는 시스템 자원의 사용을 최소화하는 제스쳐 인식시스템을 설계하고 구현한다. 이 시스템은 동적테이블 마스킹을 이용하여 노이즈를 제거하고, 가이드라인 인식 방법을 적용하여 손동작 제스쳐를 인식한다, 또한 안면 비율 분할 방법과 음영 측정 방법을 이용하여 눈과 입술의 제스쳐를 인식한다.

  • PDF

Face Detection and Region Refinement using a CNN Model (CNN 모델을 이용한 얼굴 추출 및 보정 기법)

  • Cho Il-Gook;Kim Ho-Joon
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2006.06b
    • /
    • pp.313-315
    • /
    • 2006
  • 본 연구에서는 실내에서 입력받은 영상의 조명과 크기 변화 등에 강인한 얼굴 검출 기법을 소개한다. 제안된 얼굴 검출 기법은 후보 영역 선정 과정과 얼굴패턴 검출 과정, 얼굴 영역 보정 과정으로 이루어진다. 후보 영역 선정 과정에서는 조명보정과 색상 필터, 움직임 필터를 이용하여 얼굴패턴의 후보 영역을 선정한다. 얼굴패턴 검출 과정에서는 CNN을 이용하여 특징을 추출하고, WFMM 신경망을 이용하여 얼굴 패턴을 검증한다. 얼굴 영역 보정 과정은 형태학적 연산 등의 영상 처리를 이용하여 눈 영역과 입술 영역의 위치를 판별한 후 최종적인 얼굴 영역을 결정한다.

  • PDF

Facial Expression Algorithm For Risk Situation Recognition (얼굴 표정인식을 이용한 위험상황 인지)

  • Kwak, Nae-jong;Song, Teuk-Seob
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2014.10a
    • /
    • pp.197-200
    • /
    • 2014
  • This paper proposes an algorithm for risk situation recognition using facial expression. The proposed method recognitions the surprise and fear expression among human's various emotional expression for recognizing risk situation. The proposed method firstly extracts the facial region from input, detects eye region and lip region from the extracted face. And then, the method applies Uniform LBP to each region, discriminates facial expression, and recognizes risk situation. The proposed method is evaluated for Cohn-Kanade database image. The proposed method produces good results of facial expression and discriminates risk situation well.

  • PDF

Robust Feature Extraction Based on Image-based Approach for Visual Speech Recognition (시각 음성인식을 위한 영상 기반 접근방법에 기반한 강인한 시각 특징 파라미터의 추출 방법)

  • Gyu, Song-Min;Pham, Thanh Trung;Min, So-Hee;Kim, Jing-Young;Na, Seung-You;Hwang, Sung-Taek
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.20 no.3
    • /
    • pp.348-355
    • /
    • 2010
  • In spite of development in speech recognition technology, speech recognition under noisy environment is still a difficult task. To solve this problem, Researchers has been proposed different methods where they have been used visual information except audio information for visual speech recognition. However, visual information also has visual noises as well as the noises of audio information, and this visual noises cause degradation in visual speech recognition. Therefore, it is one the field of interest how to extract visual features parameter for enhancing visual speech recognition performance. In this paper, we propose a method for visual feature parameter extraction based on image-base approach for enhancing recognition performance of the HMM based visual speech recognizer. For experiments, we have constructed Audio-visual database which is consisted with 105 speackers and each speaker has uttered 62 words. We have applied histogram matching, lip folding, RASTA filtering, Liner Mask, DCT and PCA. The experimental results show that the recognition performance of our proposed method enhanced at about 21% than the baseline method.

A Study on Lip-reading Enhancement Using Time-domain Filter (시간영역 필터를 이용한 립리딩 성능향상에 관한 연구)

  • 신도성;김진영;최승호
    • The Journal of the Acoustical Society of Korea
    • /
    • v.22 no.5
    • /
    • pp.375-382
    • /
    • 2003
  • Lip-reading technique based on bimodal is to enhance speech recognition rate in noisy environment. It is most important to detect the correct lip-image. But it is hard to estimate stable performance in dynamic environment, because of many factors to deteriorate Lip-reading's performance. There are illumination change, speaker's pronunciation habit, versatility of lips shape and rotation or size change of lips etc. In this paper, we propose the IIR filtering in time-domain for the stable performance. It is very proper to remove the noise of speech, to enhance performance of recognition by digital filtering in time domain. While the lip-reading technique in whole lip image makes data massive, the Principal Component Analysis of pre-process allows to reduce the data quantify by detection of feature without loss of image information. For the observation performance of speech recognition using only image information, we made an experiment on recognition after choosing 22 words in available car service. We used Hidden Markov Model by speech recognition algorithm to compare this words' recognition performance. As a result, while the recognition rate of lip-reading using PCA is 64%, Time-domain filter applied to lip-reading enhances recognition rate of 72.4%.

Robot vision system for face recognition using fuzzy inference from color-image (로봇의 시각시스템을 위한 칼라영상에서 퍼지추론을 이용한 얼굴인식)

  • Lee, Joo-shin
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.7 no.2
    • /
    • pp.106-110
    • /
    • 2014
  • This paper proposed the face recognition method which can be effectively applied to the robot's vision system. The proposed algorithm is recognition using hue extraction and feature point. hue extraction was using difference of skin color, pupil color, lips color. Features information were extraction from eye, nose and mouth using feature parameters of the difference between the feature point, distance ratio, angle, area. Feature parameters fuzzified data with the data generated by membership function, then evaluate the degree of similarity was the face recognition. The result of experiment are conducted with frontal color images of face as input images the received recognition rate of 96%.

A Virtual Makeup Program Using Facial Feature Area Extraction Based on Active Shape Model and Modified Alpha Blending (ASM 기반의 얼굴 특징 영역 추출 및 변형된 알파 블렌딩을 이용한 가상 메이크업 프로그램)

  • Koo, Ja-Myoung;Cho, Tai-Hoon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.14 no.8
    • /
    • pp.1827-1835
    • /
    • 2010
  • In this paper, facial feature areas in user picture are created by facial feature points extracted by ASM(Active Shape Model). In a existing virtual make-up application, users manually select a few features that are exactly. Users are uncomfortable with this method. We propose a virtual makeup application using ASM that does not require user input. In order to express a natural makeup, the modified alpha blendings for each cosmetic are used to blend skin color with cosmetic color. The Virtual makeup application was implemented to apply Foundation, Blush, Lip Stick, Lip Liner, Eye Pencil, Eye Liner and Eye Shadow.

3D Facial Synthesis and Animation for Facial Motion Estimation (얼굴의 움직임 추적에 따른 3차원 얼굴 합성 및 애니메이션)

  • Park, Do-Young;Shim, Youn-Sook;Byun, Hye-Ran
    • Journal of KIISE:Software and Applications
    • /
    • v.27 no.6
    • /
    • pp.618-631
    • /
    • 2000
  • In this paper, we suggest the method of 3D facial synthesis using the motion of 2D facial images. We use the optical flow-based method for estimation of motion. We extract parameterized motion vectors using optical flow between two adjacent image sequences in order to estimate the facial features and the facial motion in 2D image sequences. Then, we combine parameters of the parameterized motion vectors and estimate facial motion information. We use the parameterized vector model according to the facial features. Our motion vector models are eye area, lip-eyebrow area, and face area. Combining 2D facial motion information with 3D facial model action unit, we synthesize the 3D facial model.

  • PDF

Extraction of Lip Region using Chromaticity Transformation and Fuzzy Clustering (색도 변환과 퍼지 클러스터링을 이용한 입술영역 추출)

  • Kim, Jeong Yeop
    • Journal of Korea Multimedia Society
    • /
    • v.17 no.7
    • /
    • pp.806-817
    • /
    • 2014
  • The extraction of lip region is essential to Lip Reading, which is a field of image processing to get some meaningful information by the analysis of lip movement from human face image. Many conventional methods to extract lip region are proposed. One is getting the position of lip by using geometric face structure. The other discriminates lip and skin regions by using color information only. The former is more complex than the latter, however it can analyze black and white image also. The latter is very simple compared to the former, however it is very difficult to discriminate lip and skin regions because of close similarity between these two regions. And also, the accuracy is relatively low compared to the former. Conventional analysis of color coordinate systems are mostly based on specific extraction scheme for lip regions rather than coordinate system itself. In this paper, the method for selection of effective color coordinate system and chromaticity transformation to discriminate these two lip and skin region are proposed.

Korean Phonological Viseme for Lip Synch Based on Phoneme Recognition (음소인식 기반의 립싱크 구현을 위한 한국어 음운학적 Viseme의 제안)

  • Joo Heeyeol;Kang Sunmee;Ko Hanseok
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • spring
    • /
    • pp.70-73
    • /
    • 1999
  • 본 논문에서는 한국어에 대한 실시간 음소 인식을 통한 Lip Synch 구현에 필수요소인 Viseme(Visual Phoneme)을 한국어의 음운학적 접근 방법을 통해 제시하고, Lip Synch에서 입술의 모양에 결정적인 영향을 미치는 모음에 대한 모음 인식 실험 및 결과 분석을 한다.모음인식 실험에서는 한국어 음소 51개 각각에 대해 3개의 State로 이루어진 CHMM (Continilous Hidden Makov Model)으로 모델링하고, 각각의 음소가 병렬로 연결되어진 음소네트워크를 사용한다. 입력된 음성은 12차 MFCC로 특징을 추출하고, Viterbi 알고리즘을 인식 알고리즘으로 사용했으며, 인식과정에서 Bigrim 문법과 유사한 구조의 음소배열 규칙을 사용해서 인식률과 인식 속도를 향상시켰다.

  • PDF