• Title/Summary/Keyword: facial recognition

Search Result 714, Processing Time 0.022 seconds

Greedy Learning of Sparse Eigenfaces for Face Recognition and Tracking

  • Kim, Minyoung
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.14 no.3
    • /
    • pp.162-170
    • /
    • 2014
  • Appearance-based subspace models such as eigenfaces have been widely recognized as one of the most successful approaches to face recognition and tracking. The success of eigenfaces mainly has its origins in the benefits offered by principal component analysis (PCA), the representational power of the underlying generative process for high-dimensional noisy facial image data. The sparse extension of PCA (SPCA) has recently received significant attention in the research community. SPCA functions by imposing sparseness constraints on the eigenvectors, a technique that has been shown to yield more robust solutions in many applications. However, when SPCA is applied to facial images, the time and space complexity of PCA learning becomes a critical issue (e.g., real-time tracking). In this paper, we propose a very fast and scalable greedy forward selection algorithm for SPCA. Unlike a recent semidefinite program-relaxation method that suffers from complex optimization, our approach can process several thousands of data dimensions in reasonable time with little accuracy loss. The effectiveness of our proposed method was demonstrated on real-world face recognition and tracking datasets.

Design of an IOT System based on Face Recognition Technology using ESP32-CAM

  • Mahmoud, Ines;Saidi, Imen;bouzazi, Chadi
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.8
    • /
    • pp.1-6
    • /
    • 2022
  • In this paper, we will present the realization of a facial recognition system using the ESP32-CAM board controlled by an Arduino board. The goal is to monitor a remote location in real time via a camera that is integrated into the ESP32 IOT board. The acquired images will be recorded on a memory card and at the same time transmitted to a pc (a web server). The development of this remote monitoring system is to create an alternative between security, reception, and transmission of information to act accordingly. The simulation results of our proposed application of the facial recognition domain are very efficient and satisfying in real time.

Development of Facial Recognition-enabled Automatic Login for Recruitment Websites (얼굴인식을 통한 자동로그인이 가능한 채용 웹사이트 개발)

  • Hyo Hyun Choi;Min-Ho Cho
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2024.01a
    • /
    • pp.303-304
    • /
    • 2024
  • 본 논문에서는 해당 웹 사이트에 얼굴 인식을 통한 자동로그인 시스템 서비스를 구현한다. 얼굴 인식 라이브러리로 face_recognition을 사용한다. 웹 사이트에 접근 시 원하는 채용공고에 지원할 수 있으며. 원하는 기업을 검색하고 저장하여 모아 볼 수 있는 서비스를 제공하는 웹 애플리케이션을 설계하고 구현한다. React를 사용하여 프론트엔드를 구성하고 SpringBoot와 Flask를 사용하여 벡엔드를 구현하였다. 자동로그인을 위한 얼굴 인식 라이브러리로 face_recognition을 사용한다.

  • PDF

Intelligent Wheelchair System using Face and Mouth Recognition (얼굴과 입 모양 인식을 이용한 지능형 휠체어 시스템)

  • Ju, Jin-Sun;Shin, Yun-Hee;Kim, Eun-Yi
    • Journal of KIISE:Software and Applications
    • /
    • v.36 no.2
    • /
    • pp.161-168
    • /
    • 2009
  • In this paper, we develop an Intelligent Wheelchair(IW) control system for the people with various disabilities. The aim of the proposed system is to increase the mobility of severely handicapped people by providing an adaptable and effective interface for a power wheelchair. To facilitate a wide variety of user abilities, the proposed system involves the use of face-inclination and mouth-shape information, where the direction of an Intelligent Wheelchair(IW) is determined by the inclination of the user's face, while proceeding and stopping are determined by the shape of the user's mouth. To analyze these gestures, our system consists of facial feature detector, facial feature recognizer, and converter. In the stage of facial feature detector, the facial region of the intended user is first obtained using Adaboost, thereafter the mouth region detected based on edge information. The extracted features are sent to the facial feature recognizer, which recognize the face inclination and mouth shape using statistical analysis and K-means clustering, respectively. These recognition results are then delivered to a converter to control the wheelchair. When assessing the effectiveness of the proposed system with 34 users unable to utilize a standard joystick, the results showed that the proposed system provided a friendly and convenient interface.

Face Recognition using Extended Center-Symmetric Pattern and 2D-PCA (Extended Center-Symmetric Pattern과 2D-PCA를 이용한 얼굴인식)

  • Lee, Hyeon Gu;Kim, Dong Ju
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.9 no.2
    • /
    • pp.111-119
    • /
    • 2013
  • Face recognition has recently become one of the most popular research areas in the fields of computer vision, machine learning, and pattern recognition because it spans numerous applications, such as access control, surveillance, security, credit-card verification, and criminal identification. In this paper, we propose a simple descriptor called an ECSP(Extended Center-Symmetric Pattern) for illumination-robust face recognition. The ECSP operator encodes the texture information of a local face region by emphasizing diagonal components of a previous CS-LBP(Center-Symmetric Local Binary Pattern). Here, the diagonal components are emphasized because facial textures along the diagonal direction contain much more information than those of other directions. The facial texture information of the ECSP operator is then used as the input image of an image covariance-based feature extraction algorithm such as 2D-PCA(Two-Dimensional Principal Component Analysis). Performance evaluation of the proposed approach was carried out using various binary pattern operators and recognition algorithms on the Yale B database. The experimental results demonstrated that the proposed approach achieved better recognition accuracy than other approaches, and we confirmed that the proposed approach is effective against illumination variation.

Design of Face Recognition System Based on Pose Estimation : Comparative Studies of Pose Estimation Algorithms (포즈 추정 기반 얼굴 인식 시스템 설계 : 포즈 추정 알고리즘 비교 연구)

  • Kim, Jin-Yul;Kim, Jong-Bum;Oh, Sung-Kwun
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.66 no.4
    • /
    • pp.672-681
    • /
    • 2017
  • This paper is concerned with the design methodology of face recognition system based on pose estimation. In 2-dimensional face recognition, the variations of facial pose cause the deterioration of recognition performance because object recognition is carried out by using brightness of each pixel on image. To alleviate such problem, the proposed face recognition system deals with Learning Vector Quantizatioin(LVQ) or K-Nearest Neighbor(K-NN) to estimate facial pose on image and then the images obtained from LVQ or K-NN are used as the inputs of networks such as Convolution Neural Networks(CNNs) and Radial Basis Function Neural Networks(RBFNNs). The effectiveness and efficiency of the post estimation using LVQ and K-NN as well as face recognition rate using CNNs and RBFNNs are discussed through experiments carried out by using ICPR and CMU PIE databases.

Parallel Multi-task Cascade Convolution Neural Network Optimization Algorithm for Real-time Dynamic Face Recognition

  • Jiang, Bin;Ren, Qiang;Dai, Fei;Zhou, Tian;Gui, Guan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.10
    • /
    • pp.4117-4135
    • /
    • 2020
  • Due to the angle of view, illumination and scene diversity, real-time dynamic face detection and recognition is no small difficulty in those unrestricted environments. In this study, we used the intrinsic correlation between detection and calibration, using a multi-task cascaded convolutional neural network(MTCNN) to improve the efficiency of face recognition, and the output of each core network is mapped in parallel to a compact Euclidean space, where distance represents the similarity of facial features, so that the target face can be identified as quickly as possible, without waiting for all network iteration calculations to complete the recognition results. And after the angle of the target face and the illumination change, the correlation between the recognition results can be well obtained. In the actual application scenario, we use a multi-camera real-time monitoring system to perform face matching and recognition using successive frames acquired from different angles. The effectiveness of the method was verified by several real-time monitoring experiments, and good results were obtained.

Geometrical Feature-Based Detection of Pure Facial Regions (기하학적 특징에 기반한 순수 얼굴영역 검출기법)

  • 이대호;박영태
    • Journal of KIISE:Software and Applications
    • /
    • v.30 no.7_8
    • /
    • pp.773-779
    • /
    • 2003
  • Locating exact position of facial components is a key preprocessing for realizing highly accurate and reliable face recognition schemes. In this paper, we propose a simple but powerful method for detecting isolated facial components such as eyebrows, eyes, and a mouth, which are horizontally oriented and have relatively dark gray levels. The method is based on the shape-resolving locally optimum thresholding that may guarantee isolated detection of each component. We show that pure facial regions can be determined by grouping facial features satisfying simple geometric constraints on unique facial structure. In the test for over 1000 images in the AR -face database, pure facial regions were detected correctly for each face image without wearing glasses. Very few errors occurred in the face images wearing glasses with a thick frame because of the occluded eyebrow -pairs. The proposed scheme may be best suited for the later stage of classification using either the mappings or a template matching, because of its capability of handling rotational and translational variations.

Reconstruction of High-Resolution Facial Image Based on Recursive Error Back-Projection of Top-Down Machine Learning (하향식 기계학습의 반복적 오차 역투영에 기반한 고해상도 얼굴 영상의 복원)

  • Park, Jeong-Seon;Lee, Seong-Whan
    • Journal of KIISE:Software and Applications
    • /
    • v.34 no.3
    • /
    • pp.266-274
    • /
    • 2007
  • This paper proposes a new reconstruction method of high-resolution facial image from a low-resolution facial image based on top-down machine learning and recursive error back-projection. A face is represented by a linear combination of prototypes of shape and that of texture. With the shape and texture information of each pixel in a given low-resolution facial image, we can estimate optimal coefficients for a linear combination of prototypes of shape and those that of texture by solving least square minimizations. Then high-resolution facial image can be obtained by using the optimal coefficients for linear combination of the high-resolution prototypes. In addition, a recursive error back-projection procedure is applied to improve the reconstruction accuracy of high-resolution facial image. The encouraging results of the proposed method show that our method can be used to improve the performance of the face recognition by applying our method to reconstruct high-resolution facial images from low-resolution images captured at a distance.

Dynamic Facial Expression of Fuzzy Modeling Using Probability of Emotion (감정확률을 이용한 동적 얼굴표정의 퍼지 모델링)

  • Kang, Hyo-Seok;Baek, Jae-Ho;Kim, Eun-Tai;Park, Mignon
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.19 no.1
    • /
    • pp.1-5
    • /
    • 2009
  • This paper suggests to apply mirror-reflected method based 2D emotion recognition database to 3D application. Also, it makes facial expression of fuzzy modeling using probability of emotion. Suggested facial expression function applies fuzzy theory to 3 basic movement for facial expressions. This method applies 3D application to feature vector for emotion recognition from 2D application using mirror-reflected multi-image. Thus, we can have model based on fuzzy nonlinear facial expression of a 2D model for a real model. We use average values about probability of 6 basic expressions such as happy, sad, disgust, angry, surprise and fear. Furthermore, dynimic facial expressions are made via fuzzy modelling. This paper compares and analyzes feature vectors of real model with 3D human-like avatar.