• Title/Summary/Keyword: face images

Search Result 995, Processing Time 0.024 seconds

Collaborative Local Active Appearance Models for Illuminated Face Images (조명얼굴 영상을 위한 협력적 지역 능동표현 모델)

  • Yang, Jun-Young;Ko, Jae-Pil;Byun, Hye-Ran
    • Journal of KIISE:Software and Applications
    • /
    • v.36 no.10
    • /
    • pp.816-824
    • /
    • 2009
  • In the face space, face images due to illumination and pose variations have a nonlinear distribution. Active Appearance Models (AAM) based on the linear model have limits to the nonlinear distribution of face images. In this paper, we assume that a few clusters of face images are given; we build local AAMs according to the clusters of face images, and then select a proper AAM model during the fitting phase. To solve the problem of updating fitting parameters among the models due to the model changing, we propose to build in advance relationships among the clusters in the parameter space from the training images. In addition, we suggest a gradual model changing to reduce improper model selections due to serious fitting failures. In our experiment, we apply the proposed model to Yale Face Database B and compare it with the previous method. The proposed method demonstrated successful fitting results with strongly illuminated face images of deep shadows.

FD-StackGAN: Face De-occlusion Using Stacked Generative Adversarial Networks

  • Jabbar, Abdul;Li, Xi;Iqbal, M. Munawwar;Malik, Arif Jamal
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.7
    • /
    • pp.2547-2567
    • /
    • 2021
  • It has been widely acknowledged that occlusion impairments adversely distress many face recognition algorithms' performance. Therefore, it is crucial to solving the problem of face image occlusion in face recognition. To solve the image occlusion problem in face recognition, this paper aims to automatically de-occlude the human face majority or discriminative regions to improve face recognition performance. To achieve this, we decompose the generative process into two key stages and employ a separate generative adversarial network (GAN)-based network in both stages. The first stage generates an initial coarse face image without an occlusion mask. The second stage refines the result from the first stage by forcing it closer to real face images or ground truth. To increase the performance and minimize the artifacts in the generated result, a new refine loss (e.g., reconstruction loss, perceptual loss, and adversarial loss) is used to determine all differences between the generated de-occluded face image and ground truth. Furthermore, we build occluded face images and corresponding occlusion-free face images dataset. We trained our model on this new dataset and later tested it on real-world face images. The experiment results (qualitative and quantitative) and the comparative study confirm the robustness and effectiveness of the proposed work in removing challenging occlusion masks with various structures, sizes, shapes, types, and positions.

Face region detection algorithm of natural-image (자연 영상에서 얼굴영역 검출 알고리즘)

  • Lee, Joo-shin
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.7 no.1
    • /
    • pp.55-60
    • /
    • 2014
  • In this paper, we proposed a method for face region extraction by skin-color hue, saturation and facial feature extraction in natural images. The proposed algorithm is composed of lighting correction and face detection process. In the lighting correction step, performing correction function for a lighting change. The face detection process extracts the area of skin color by calculating Euclidian distances to the input images using as characteristic vectors color and chroma in 20 skin color sample images. Eye detection using C element in the CMY color model and mouth detection using Q element in the YIQ color model for extracted candidate areas. Face area detected based on human face knowledge for extracted candidate areas. When an experiment was conducted with 10 natural images of face as input images, the method showed a face detection rate of 100%.

Facial Shape Recognition Using Self Organized Feature Map(SOFM)

  • Kim, Seung-Jae;Lee, Jung-Jae
    • International journal of advanced smart convergence
    • /
    • v.8 no.4
    • /
    • pp.104-112
    • /
    • 2019
  • This study proposed a robust detection algorithm. It detects face more stably with respect to changes in light and rotation forthe identification of a face shape. The proposed algorithm uses face shape asinput information in a single camera environment and divides only face area through preprocessing process. However, it is not easy to accurately recognize the face area that is sensitive to lighting changes and has a large degree of freedom, and the error range is large. In this paper, we separated the background and face area using the brightness difference of the two images to increase the recognition rate. The brightness difference between the two images means the difference between the images taken under the bright light and the images taken under the dark light. After separating only the face region, the face shape is recognized by using the self-organization feature map (SOFM) algorithm. SOFM first selects the first top neuron through the learning process. Second, the highest neuron is renewed by competing again between the highest neuron and neighboring neurons through the competition process. Third, the final top neuron is selected by repeating the learning process and the competition process. In addition, the competition will go through a three-step learning process to ensure that the top neurons are updated well among neurons. By using these SOFM neural network algorithms, we intend to implement a stable and robust real-time face shape recognition system in face shape recognition.

Face Recognition Research Based on Multi-Layers Residual Unit CNN Model

  • Zhang, Ruyang;Lee, Eung-Joo
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.11
    • /
    • pp.1582-1590
    • /
    • 2022
  • Due to the situation of the widespread of the coronavirus, which causes the problem of lack of face image data occluded by masks at recent time, in order to solve the related problems, this paper proposes a method to generate face images with masks using a combination of generative adversarial networks and spatial transformation networks based on CNN model. The system we proposed in this paper is based on the GAN, combined with multi-scale convolution kernels to extract features at different details of the human face images, and used Wasserstein divergence as the measure of the distance between real samples and synthetic samples in order to optimize Generator performance. Experiments show that the proposed method can effectively put masks on face images with high efficiency and fast reaction time and the synthesized human face images are pretty natural and real.

A Secure Face Cryptogr aphy for Identity Document Based on Distance Measures

  • Arshad, Nasim;Moon, Kwang-Seok;Kim, Jong-Nam
    • Journal of Korea Multimedia Society
    • /
    • v.16 no.10
    • /
    • pp.1156-1162
    • /
    • 2013
  • Face verification has been widely studied during the past two decades. One of the challenges is the rising concern about the security and privacy of the template database. In this paper, we propose a secure face verification system which generates a unique secure cryptographic key from a face template. The face images are processed to produce face templates or codes to be utilized for the encryption and decryption tasks. The result identity data is encrypted using Advanced Encryption Standard (AES). Distance metric naming hamming distance and Euclidean distance are used for template matching identification process, where template matching is a process used in pattern recognition. The proposed system is tested on the ORL, YALEs, and PKNU face databases, which contain 360, 135, and 54 training images respectively. We employ Principle Component Analysis (PCA) to determine the most discriminating features among face images. The experimental results showed that the proposed distance measure was one the promising best measures with respect to different characteristics of the biometric systems. Using the proposed method we needed to extract fewer images in order to achieve 100% cumulative recognition than using any other tested distance measure.

Face Recognition using Emotional Face Images and Fuzzy Fisherface (감정이 있는 얼굴영상과 퍼지 Fisherface를 이용한 얼굴인식)

  • Koh, Hyun-Joo;Chun, Myung-Geun;Paliwal, K.K.
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.15 no.1
    • /
    • pp.94-98
    • /
    • 2009
  • In this paper, we deal with a face recognition method for the emotional face images. Since the face recognition is one of the most natural and straightforward biometric methods, there have been various research works. However, most of them are focused on the expressionless face images and have had a very difficult problem if we consider the facial expression. In real situations, however, it is required to consider the emotional face images. Here, three basic human emotions such as happiness, sadness, and anger are investigated for the face recognition. And, this situation requires a robust face recognition algorithm then we use a fuzzy Fisher's Linear Discriminant (FLD) algorithm with the wavelet transform. The fuzzy Fisherface is a statistical method that maximizes the ratio of between-scatter matrix and within-scatter matrix and also handles the fuzzy class information. The experimental results obtained for the CBNU face databases reveal that the approach presented in this paper yields better recognition performance in comparison with the results obtained by other recognition methods.

DETECTION OF FACIAL FEATURES IN COLOR IMAGES WITH VARIOUS BACKGROUNDS AND FACE POSES

  • Park, Jae-Young;Kim, Nak-Bin
    • Journal of Korea Multimedia Society
    • /
    • v.6 no.4
    • /
    • pp.594-600
    • /
    • 2003
  • In this paper, we propose a detection method for facial features in color images with various backgrounds and face poses. To begin with, the proposed method extracts face candidacy region from images with various backgrounds, which have skin-tone color and complex objects, via the color and edge information of face. And then, by using the elliptical shape property of face, we correct a rotation, scale, and tilt of face region caused by various poses of head. Finally, we verify the face using features of face and detect facial features. In our experimental results, it is shown that accuracy of detection is high and the proposed method can be used in pose-invariant face recognition system effectively

  • PDF

Untact Face Recognition System Based on Super-resolution in Low-Resolution Images (초고해상도 기반 비대면 저해상도 영상의 얼굴 인식 시스템)

  • Bae, Hyeon Bin;Kwon, Oh Seol
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.3
    • /
    • pp.412-420
    • /
    • 2020
  • This paper proposes a performance-improving face recognition system based on a super resolution method for low-resolution images. The conventional face recognition algorithm has a rapidly decreased accuracy rate due to small image resolution by a distance. To solve the previously mentioned problem, this paper generates a super resolution images based o deep learning method. The proposed method improved feature information from low-resolution images using a super resolution method and also applied face recognition using a feature extraction and an classifier. In experiments, the proposed method improves the face recognition rate when compared to conventional methods.

Face recognition by using independent component analysis (독립 성분 분석을 이용한 얼굴인식)

  • 김종규;장주석;김영일
    • Journal of the Korean Institute of Telematics and Electronics C
    • /
    • v.35C no.10
    • /
    • pp.48-58
    • /
    • 1998
  • We present a method that can recognize face images using independent component analysis that is used mainly for blind sources separation in signal processing. We assumed that a face image can be expressed as the sum of a set of statistically independent feature images, which was obtained by using independent component analysis. Face recognition was peformed by projecting the input image to the feature image space and then by comparing its projection components with those of stored reference images. We carried out face recognition experiments with a database that consists of various varied face images (total 400 varied facial images collected from 10 per person) and compared the performance of our method with that of the eigenface method based on principal component analysis. The presented method gave better results of recognition rate than the eigenface method did, and showed robustness to the random noise added in the input facial images.

  • PDF