• Title/Summary/Keyword: Video Face Recognition

Search Result 110, Processing Time 0.025 seconds

Low Resolution Rate Face Recognition Based on Multi-scale CNN

  • Wang, Ji-Yuan;Lee, Eung-Joo
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.12
    • /
    • pp.1467-1472
    • /
    • 2018
  • For the problem that the face image of surveillance video cannot be accurately identified due to the low resolution, this paper proposes a low resolution face recognition solution based on convolutional neural network model. Convolutional Neural Networks (CNN) model for multi-scale input The CNN model for multi-scale input is an improvement over the existing "two-step method" in which low-resolution images are up-sampled using a simple bi-cubic interpolation method. Then, the up sampled image and the high-resolution image are mixed as a model training sample. The CNN model learns the common feature space of the high- and low-resolution images, and then measures the feature similarity through the cosine distance. Finally, the recognition result is given. The experiments on the CMU PIE and Extended Yale B datasets show that the accuracy of the model is better than other comparison methods. Compared with the CMDA_BGE algorithm with the highest recognition rate, the accuracy rate is 2.5%~9.9%.

Multimodal Biometrics Recognition from Facial Video with Missing Modalities Using Deep Learning

  • Maity, Sayan;Abdel-Mottaleb, Mohamed;Asfour, Shihab S.
    • Journal of Information Processing Systems
    • /
    • v.16 no.1
    • /
    • pp.6-29
    • /
    • 2020
  • Biometrics identification using multiple modalities has attracted the attention of many researchers as it produces more robust and trustworthy results than single modality biometrics. In this paper, we present a novel multimodal recognition system that trains a deep learning network to automatically learn features after extracting multiple biometric modalities from a single data source, i.e., facial video clips. Utilizing different modalities, i.e., left ear, left profile face, frontal face, right profile face, and right ear, present in the facial video clips, we train supervised denoising auto-encoders to automatically extract robust and non-redundant features. The automatically learned features are then used to train modality specific sparse classifiers to perform the multimodal recognition. Moreover, the proposed technique has proven robust when some of the above modalities were missing during the testing. The proposed system has three main components that are responsible for detection, which consists of modality specific detectors to automatically detect images of different modalities present in facial video clips; feature selection, which uses supervised denoising sparse auto-encoders network to capture discriminative representations that are robust to the illumination and pose variations; and classification, which consists of a set of modality specific sparse representation classifiers for unimodal recognition, followed by score level fusion of the recognition results of the available modalities. Experiments conducted on the constrained facial video dataset (WVU) and the unconstrained facial video dataset (HONDA/UCSD), resulted in a 99.17% and 97.14% Rank-1 recognition rates, respectively. The multimodal recognition accuracy demonstrates the superiority and robustness of the proposed approach irrespective of the illumination, non-planar movement, and pose variations present in the video clips even in the situation of missing modalities.

Automatic Cast-list Analysis System in Broadcasting Videos (방송 비디오 등장인물 자동 분석 시스템)

  • 김기남;김형준;김회율
    • Journal of Broadcast Engineering
    • /
    • v.9 no.2
    • /
    • pp.164-173
    • /
    • 2004
  • In this paper, we propose a system that can analyze appearance interval of casts by detecting and recognizing casts in broadcasting videos. The cast is one of the most important characteristics in broadcasting videos such as drama and sports. In this paper, we propose the ACAV(Automatic Cast-list Analysis in Videos) system that analyzes cast-list automatically in video. The ACAV system consists of FAGIS(FAce reGIStration) which registers detected faces into the face DB and FACOG(FAce reCOGnition) that analyses the cast-list in video sequence using the face DB. We evaluate performance of the ACAV system by comparing with FaceIt, one of the most well-known commercial systems for the cast-list analysis. The ACAV shows face detection and recognition rates of 84.3% and 75.7% that are about 30% and 27.5% higher than those of FaceIt, respectively. The ACAV system can be applied to mass broadcasting videos management system for broadcasters and video management system of PVR(Personal Video Recorder) and mobile phone for the public.

A Search Model Using Time Interval Variation to Identify Face Recognition Results

  • Choi, Yun-seok;Lee, Wan Yeon
    • International journal of advanced smart convergence
    • /
    • v.11 no.3
    • /
    • pp.64-71
    • /
    • 2022
  • Various types of attendance management systems are being introduced in a remote working environment and research on using face recognition is in progress. To ensure accurate worker's attendance, a face recognition-based attendance management system must analyze every frame of video, but face recognition is a heavy task, the number of the task should be minimized without affecting accuracy. In this paper, we proposed a search model using time interval variation to minimize the number of face recognition task of recorded videos for attendance management system. The proposed model performs face recognition by changing the interval of the frame identification time when there is no change in the attendance status for a certain period. When a change in the face recognition status occurs, it moves in the reverse direction and performs frame checks to more accurate attendance time checking. The implementation of proposed model performed at least 4.5 times faster than all frame identification and showed at least 97% accuracy.

Object Recognition Face Detection With 3D Imaging Parameters A Research on Measurement Technology (3D영상 객체인식을 통한 얼굴검출 파라미터 측정기술에 대한 연구)

  • Choi, Byung-Kwan;Moon, Nam-Mee
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.10
    • /
    • pp.53-62
    • /
    • 2011
  • In this paper, high-tech IT Convergence, to the development of complex technology, special technology, video object recognition technology was considered only as a smart - phone technology with the development of personal portable terminal has been developed crossroads. Technology-based detection of 3D face recognition technology that recognizes objects detected through the intelligent video recognition technology has been evolving technologies based on image recognition, face detection technology with through the development speed is booming. In this paper, based on human face recognition technology to detect the object recognition image processing technology is applied through the face recognition technology applied to the IP camera is the party of the mouth, and allowed the ability to identify and apply the human face recognition, measurement techniques applied research is suggested. Study plan: 1) face model based face tracking technology was developed and applied 2) algorithm developed by PC-based measurement of human perception through the CPU load in the face value of their basic parameters can be tracked, and 3) bilateral distance and the angle of gaze can be tracked in real time, proved effective.

Improving Indentification Performance by Integrating Evidence From Evidence

  • Park, Kwang-Chae;Kim, Young-Geil;Cheong, Ha-Young
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.9 no.6
    • /
    • pp.546-552
    • /
    • 2016
  • We present a quantitative evaluation of an algorithm for model-based face recognition. The algorithm actively learns how individual faces vary through video sequences, providing on-line suppression of confounding factors such as expression, lighting and pose. By actively decoupling sources of image variation, the algorithm provides a framework in which identity evidence can be integrated over a sequence. We demonstrate that face recognition can be considerably improved by the analysis of video sequences. The method presented is widely applicable in many multi-class interpretation problems.

Comparison of satisfaction, interest, and experience awareness of 360° virtual reality video and first-person video in non-face-to-face practical lectures in medical emergency departments (응급구조학과 비대면 실습 강의에서 360° 가상현실 영상과 1인칭 시점 영상의 만족도, 흥미도, 경험인식 비교)

  • Lee, Hyo-Ju;Shin, Sang-Yol;Jung, Eun-Kyung
    • The Korean Journal of Emergency Medical Services
    • /
    • v.24 no.3
    • /
    • pp.55-63
    • /
    • 2020
  • Purpose: This study aimed to establish effective training strategies and methods by comparing the effects of 360° virtual reality video and first-person video in non-face-to-face practical lectures. Methods: This crossover study, implemented May 18-31, 2020, included 27 participants. We compared 360° virtual reality video and first-person video. SPSS version 25.0 was used for statistical analysis. Results: The 360° virtual reality video had a higher score of experience recognition (p=.039), vividness (p=.045), presence (p=.000), fantasy factor (p=.000) than the first-person video, but no significant difference was indicated for satisfaction (p=.348) or interest (p=.441). Conclusion: 360° virtual reality video and first-person video can be used as training alternatives to achieve the standard educational objectives in non-face-to-face practical lectures.

A Video Expression Recognition Method Based on Multi-mode Convolution Neural Network and Multiplicative Feature Fusion

  • Ren, Qun
    • Journal of Information Processing Systems
    • /
    • v.17 no.3
    • /
    • pp.556-570
    • /
    • 2021
  • The existing video expression recognition methods mainly focus on the spatial feature extraction of video expression images, but tend to ignore the dynamic features of video sequences. To solve this problem, a multi-mode convolution neural network method is proposed to effectively improve the performance of facial expression recognition in video. Firstly, OpenFace 2.0 is used to detect face images in video, and two deep convolution neural networks are used to extract spatiotemporal expression features. Furthermore, spatial convolution neural network is used to extract the spatial information features of each static expression image, and the dynamic information feature is extracted from the optical flow information of multiple expression images based on temporal convolution neural network. Then, the spatiotemporal features learned by the two deep convolution neural networks are fused by multiplication. Finally, the fused features are input into support vector machine to realize the facial expression classification. Experimental results show that the recognition accuracy of the proposed method can reach 64.57% and 60.89%, respectively on RML and Baum-ls datasets. It is better than that of other contrast methods.

Face Tracking and Recognition in Video with PCA-based Pose-Classification and (2D)2PCA recognition algorithm (비디오속의 얼굴추적 및 PCA기반 얼굴포즈분류와 (2D)2PCA를 이용한 얼굴인식)

  • Kim, Jin-Yul;Kim, Yong-Seok
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.23 no.5
    • /
    • pp.423-430
    • /
    • 2013
  • In typical face recognition systems, the frontal view of face is preferred to reduce the complexity of the recognition. Thus individuals may be required to stare into the camera, or the camera should be located so that the frontal images are acquired easily. However these constraints severely restrict the adoption of face recognition to wide applications. To alleviate this problem, in this paper, we address the problem of tracking and recognizing faces in video captured with no environmental control. The face tracker extracts a sequence of the angle/size normalized face images using IVT (Incremental Visual Tracking) algorithm that is known to be robust to changes in appearance. Since no constraints have been imposed between the face direction and the video camera, there will be various poses in face images. Thus the pose is identified using a PCA (Principal Component Analysis)-based pose classifier, and only the pose-matched face images are used to identify person against the pre-built face DB with 5-poses. For face recognition, PCA, (2D)PCA, and $(2D)^2PCA$ algorithms have been tested to compute the recognition rate and the execution time.

Using a Multi-Faced Technique SPFACS Video Object Design Analysis of The AAM Algorithm Applies Smile Detection (다면기법 SPFACS 영상객체를 이용한 AAM 알고리즘 적용 미소검출 설계 분석)

  • Choi, Byungkwan
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.11 no.3
    • /
    • pp.99-112
    • /
    • 2015
  • Digital imaging technology has advanced beyond the limits of the multimedia industry IT convergence, and to develop a complex industry, particularly in the field of object recognition, face smart-phones associated with various Application technology are being actively researched. Recently, face recognition technology is evolving into an intelligent object recognition through image recognition technology, detection technology, the detection object recognition through image recognition processing techniques applied technology is applied to the IP camera through the 3D image object recognition technology Face Recognition been actively studied. In this paper, we first look at the essential human factor, technical factors and trends about the technology of the human object recognition based SPFACS(Smile Progress Facial Action Coding System)study measures the smile detection technology recognizes multi-faceted object recognition. Study Method: 1)Human cognitive skills necessary to analyze the 3D object imaging system was designed. 2)3D object recognition, face detection parameter identification and optimal measurement method using the AAM algorithm inside the proposals and 3)Face recognition objects (Face recognition Technology) to apply the result to the recognition of the person's teeth area detecting expression recognition demonstrated by the effect of extracting the feature points.