• Title/Summary/Keyword: Face expression

Search Result 454, Processing Time 0.028 seconds

Study of Facial Expression Recognition using Variable-sized Block (가변 크기 블록(Variable-sized Block)을 이용한 얼굴 표정 인식에 관한 연구)

  • Cho, Youngtak;Ryu, Byungyong;Chae, Oksam
    • Convergence Security Journal
    • /
    • v.19 no.1
    • /
    • pp.67-78
    • /
    • 2019
  • Most existing facial expression recognition methods use a uniform grid method that divides the entire facial image into uniform blocks when describing facial features. The problem of this method may include non-face backgrounds, which interferes with discrimination of facial expressions, and the feature of a face included in each block may vary depending on the position, size, and orientation of the face in the input image. In this paper, we propose a variable-size block method which determines the size and position of a block that best represents meaningful facial expression change. As a part of the effort, we propose the way to determine the optimal number, position and size of each block based on the facial feature points. For the evaluation of the proposed method, we generate the facial feature vectors using LDTP and construct a facial expression recognition system based on SVM. Experimental results show that the proposed method is superior to conventional uniform grid based method. Especially, it shows that the proposed method can adapt to the change of the input environment more effectively by showing relatively better performance than exiting methods in the images with large shape and orientation changes.

Extreme Learning Machine Ensemble Using Bagging for Facial Expression Recognition

  • Ghimire, Deepak;Lee, Joonwhoan
    • Journal of Information Processing Systems
    • /
    • v.10 no.3
    • /
    • pp.443-458
    • /
    • 2014
  • An extreme learning machine (ELM) is a recently proposed learning algorithm for a single-layer feed forward neural network. In this paper we studied the ensemble of ELM by using a bagging algorithm for facial expression recognition (FER). Facial expression analysis is widely used in the behavior interpretation of emotions, for cognitive science, and social interactions. This paper presents a method for FER based on the histogram of orientation gradient (HOG) features using an ELM ensemble. First, the HOG features were extracted from the face image by dividing it into a number of small cells. A bagging algorithm was then used to construct many different bags of training data and each of them was trained by using separate ELMs. To recognize the expression of the input face image, HOG features were fed to each trained ELM and the results were combined by using a majority voting scheme. The ELM ensemble using bagging improves the generalized capability of the network significantly. The two available datasets (JAFFE and CK+) of facial expressions were used to evaluate the performance of the proposed classification system. Even the performance of individual ELM was smaller and the ELM ensemble using a bagging algorithm improved the recognition performance significantly.

Synthesis of Expressive Talking Heads from Speech with Recurrent Neural Network (RNN을 이용한 Expressive Talking Head from Speech의 합성)

  • Sakurai, Ryuhei;Shimba, Taiki;Yamazoe, Hirotake;Lee, Joo-Ho
    • The Journal of Korea Robotics Society
    • /
    • v.13 no.1
    • /
    • pp.16-25
    • /
    • 2018
  • The talking head (TH) indicates an utterance face animation generated based on text and voice input. In this paper, we propose the generation method of TH with facial expression and intonation by speech input only. The problem of generating TH from speech can be regarded as a regression problem from the acoustic feature sequence to the facial code sequence which is a low dimensional vector representation that can efficiently encode and decode a face image. This regression was modeled by bidirectional RNN and trained by using SAVEE database of the front utterance face animation database as training data. The proposed method is able to generate TH with facial expression and intonation TH by using acoustic features such as MFCC, dynamic elements of MFCC, energy, and F0. According to the experiments, the configuration of the BLSTM layer of the first and second layers of bidirectional RNN was able to predict the face code best. For the evaluation, a questionnaire survey was conducted for 62 persons who watched TH animations, generated by the proposed method and the previous method. As a result, 77% of the respondents answered that the proposed method generated TH, which matches well with the speech.

Pose Invariant View-Based Enhanced Fisher Linear Discriminant Models for Face Recognition

  • Lee, Sung-Oh;Park, Gwi-Tae
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2001.10a
    • /
    • pp.101.2-101
    • /
    • 2001
  • This paper proposes a novel face recognition algorithm to recognize human face robustly under various conditions, such as changes of pose, illumination, and expression, etc. at indoor environments. A conventional automatic face recognition system consists of the detection and the recognition part. Generally, the detection part is dominant over the other part in the estimating whole recognition rate. So, in this paper, we suggest the view-specific eigenface method as preprocessor to estimate various poses of the face in the input image. Then, we apply the Enhanced FLD Models (EFM) to the result of it, twice. Because, the EFM recognizes human face, and reduces the error of standardization effectively. To deal with view-varying problem, we build one basis vector set for each view individually. Finally, the dimensionalities of ...

  • PDF

Skin Color Based Facial Features Extraction

  • Alom, Md. Zahangir;Lee, Hyo Jong
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2011.11a
    • /
    • pp.351-354
    • /
    • 2011
  • This paper discusses on facial features extraction based on proposed skin color model. Different parts of face from input image are segmented based on skin color model. Moreover, this paper also discusses on concept to detect the eye and mouth position on face. A height and width ratio (${\delta}=1.1618$) based technique is also proposed to accurate detection of face region from the segmented image. Finally, we have cropped the desired part of the face. This exactly exacted face part is useful for face recognition and detection, facial feature analysis and expression analysis. Experimental results of propose method shows that the proposed method is robust and accurate.

3D Facial Animation with Head Motion Estimation and Facial Expression Cloning (얼굴 모션 추정과 표정 복제에 의한 3차원 얼굴 애니메이션)

  • Kwon, Oh-Ryun;Chun, Jun-Chul
    • The KIPS Transactions:PartB
    • /
    • v.14B no.4
    • /
    • pp.311-320
    • /
    • 2007
  • This paper presents vision-based 3D facial expression animation technique and system which provide the robust 3D head pose estimation and real-time facial expression control. Many researches of 3D face animation have been done for the facial expression control itself rather than focusing on 3D head motion tracking. However, the head motion tracking is one of critical issues to be solved for developing realistic facial animation. In this research, we developed an integrated animation system that includes 3D head motion tracking and facial expression control at the same time. The proposed system consists of three major phases: face detection, 3D head motion tracking, and facial expression control. For face detection, with the non-parametric HT skin color model and template matching, we can detect the facial region efficiently from video frame. For 3D head motion tracking, we exploit the cylindrical head model that is projected to the initial head motion template. Given an initial reference template of the face image and the corresponding head motion, the cylindrical head model is created and the foil head motion is traced based on the optical flow method. For the facial expression cloning we utilize the feature-based method, The major facial feature points are detected by the geometry of information of the face with template matching and traced by optical flow. Since the locations of varying feature points are composed of head motion and facial expression information, the animation parameters which describe the variation of the facial features are acquired from geometrically transformed frontal head pose image. Finally, the facial expression cloning is done by two fitting process. The control points of the 3D model are varied applying the animation parameters to the face model, and the non-feature points around the control points are changed by use of Radial Basis Function(RBF). From the experiment, we can prove that the developed vision-based animation system can create realistic facial animation with robust head pose estimation and facial variation from input video image.

Face Recognition: A Survey (얼굴인식 기술동향)

  • Mun, Hyeon-Jun
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02c
    • /
    • pp.172-177
    • /
    • 2008
  • Biometrics is essential for person identification because of its uniqueness from each individuals. Face recognition technology has advantage over other biometrics because of its convenience and non-intrusive characteristics. In this paper, we will present a overview of face recognition technology including face detection, feature extraction, and face recognition system. For face detection, we will describe template based method and face component based approach. PCA and LDA approach will be discussed for feature extraction, and nearest neighbor classifiers -will be covered for matching. Large database and the standardized performance evaluation methodology is essential in order to support state-of-the-art face recognition system. Also, 3D based face recognition technology is the key solution for the pose, lighting and expression variations in many applications.

  • PDF

Faces of the Face

  • Choi, Jeongho
    • Archives of Plastic Surgery
    • /
    • v.44 no.3
    • /
    • pp.251-256
    • /
    • 2017
  • The most important environment of human being is the human being itself. So we have been sensitive to the appearance of ours and others at the same time. This writing aims for locating origins of the face and discerning differences [1] between the face of humans and those of other animals [2]. The face assumes a substantial significance not merely as a body function, but, above all, a means of expressions and features being looked at. The face is an important means of communication to humans as social animals. Knowledges about the various faces of the face are useful to become a efficient specialist as an extensive generalist because the face is a regular patron to the plastic surgery. The face in Korean consists of two elements of eol (the soul or the mind) and gul a residing place). When Wittgenstein says "the face is the soul of the body," his semantics corresponds to the Korean meaning. The meaning of the face in Korean is summed up in five ways. (1) the head or the front of the face with the eye, the nose and the mouth, (2) reputation or honor, (3) the general description of the psychological state, such as "the face of sadness", (4) a figure person representing a particular area, such as "Sun Dong-yul is the face of the Korean baseball community," (5) the primary imagery of the things and the event, such as "He is the face of the 4.19 Revolution." As such, the word "face", referring to a body part, extends its usages in a wide variety of contexts. What image do you convoke when you think of a person? With rare exceptions, you are most likely to invoke the face of the person. The face has come to be a byword for one's reputation or honor, and a pronoun for an expression of the essence of the thing and the event. This is presumably true of other languages. That is because human beings are equipped with the universal rule of language. A comprehensive understanding of the face is a must for cosmetic surgeons whose main responsibility is to sculpt and repair the face (Fig. 1).

A New Image Processing Scheme For Face Swapping Using CycleGAN (순환 적대적 생성 신경망을 이용한 안면 교체를 위한 새로운 이미지 처리 기법)

  • Ban, Tae-Won
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.9
    • /
    • pp.1305-1311
    • /
    • 2022
  • With the recent rapid development of mobile terminals and personal computers and the advent of neural network technology, real-time face swapping using images has become possible. In particular, the cycle generative adversarial network made it possible to replace faces using uncorrelated image data. In this paper, we propose an input data processing scheme that can improve the quality of face swapping with less training data and time. The proposed scheme can improve the image quality while preserving facial structure and expression information by combining facial landmarks extracted through a pre-trained neural network with major information that affects the structure and expression of the face. Using the blind/referenceless image spatial quality evaluator (BRISQUE) score, which is one of the AI-based non-reference quality metrics, we quantitatively analyze the performance of the proposed scheme and compare it to the conventional schemes. According to the numerical results, the proposed scheme obtained BRISQUE scores improved by about 4.6% to 14.6%, compared to the conventional schemes.

Face Recognition using Karhunen-Loeve projection and Elastic Graph Matching (Karhunen-Loeve 근사 방법과 Elastic Graph Matching을 병합한 얼굴 인식)

  • 이형지;이완수;정재호
    • Proceedings of the IEEK Conference
    • /
    • 2001.06d
    • /
    • pp.231-234
    • /
    • 2001
  • This paper proposes a face recognition technique that effectively combines elastic graph matching (EGM) and Fisherface algorithm. EGM as one of dynamic lint architecture uses not only face-shape but also the gray information of image, and Fisherface algorithm as a class specific method is robust about variations such as lighting direction and facial expression. In the proposed face recognition adopting the above two methods, the linear projection per node of an image graph reduces dimensionality of labeled graph vector and provides a feature space to be used effectively for the classification. In comparison with a conventional method, the proposed approach could obtain satisfactory results in the perspectives of recognition rates and speeds. Especially, we could get maximum recognition rate of 99.3% by leaving-one-out method for the experiments with the Yale Face Databases.

  • PDF