• Title/Summary/Keyword: Video Face Recognition

Search Result 110, Processing Time 0.028 seconds

Smart Mirror for Styling (스타일링을 위한 스마트 미러)

  • Kang, Su-Bin;Kwon, Seung-Ha;Kim, Yun-Ho;Lee, Soo-Ik;Han, Young-Oh
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.16 no.6
    • /
    • pp.1271-1278
    • /
    • 2021
  • In this paper, we implemented smart mirrors that virtually select and experience eyeglasses and hair styles on user faces through face recognition and recommend weather-specific clothes to guide various styles. In addition, makeup is possible while watching the video while looking at the screen. Raspberry Pi, acrylic plate and half mirror film were used to reduce the cost of conventional smart mirrors. It also added basic information such as weather, dates, calendars, and news, and increased user convenience by using a touchscreen.

Mask Wearing Detection System using Deep Learning (딥러닝을 이용한 마스크 착용 여부 검사 시스템)

  • Nam, Chung-hyeon;Nam, Eun-jeong;Jang, Kyung-Sik
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.1
    • /
    • pp.44-49
    • /
    • 2021
  • Recently, due to COVID-19, studies have been popularly worked to apply neural network to mask wearing automatic detection system. For applying neural networks, the 1-stage detection or 2-stage detection methods are used, and if data are not sufficiently collected, the pretrained neural network models are studied by applying fine-tuning techniques. In this paper, the system is consisted of 2-stage detection method that contain MTCNN model for face recognition and ResNet model for mask detection. The mask detector was experimented by applying five ResNet models to improve accuracy and fps in various environments. Training data used 17,217 images that collected using web crawler, and for inference, we used 1,913 images and two one-minute videos respectively. The experiment showed a high accuracy of 96.39% for images and 92.98% for video, and the speed of inference for video was 10.78fps.

Wavelet Transform-based Face Detection for Real-time Applications (실시간 응용을 위한 웨이블릿 변환 기반의 얼굴 검출)

  • 송해진;고병철;변혜란
    • Journal of KIISE:Software and Applications
    • /
    • v.30 no.9
    • /
    • pp.829-842
    • /
    • 2003
  • In this Paper, we propose the new face detection and tracking method based on template matching for real-time applications such as, teleconference, telecommunication, front stage of surveillance system using face recognition, and video-phone applications. Since the main purpose of paper is to track a face regardless of various environments, we use template-based face tracking method. To generate robust face templates, we apply wavelet transform to the average face image and extract three types of wavelet template from transformed low-resolution average face. However template matching is generally sensitive to the change of illumination conditions, we apply Min-max normalization with histogram equalization according to the variation of intensity. Tracking method is also applied to reduce the computation time and predict precise face candidate region. Finally, facial components are also detected and from the relative distance of two eyes, we estimate the size of facial ellipse.

Implementation of CoMirror System with Video Call and Messaging Function between Smart Mirrors (스마트 미러간 화상 통화와 메시징 기능을 가진 CoMirror 시스템 구현)

  • Hwang, Kitae;Kim, Kyung-Mi;Kim, Yu-Jin;Park, Chae-Won;Yoo, Song-Yeon;Jung, Inhwan;Lee, Jae-Moon
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.22 no.6
    • /
    • pp.121-127
    • /
    • 2022
  • Smart mirror is an IoT device that attaches a display and an embedded computer to the mirror and provides various information to the useer along with the mirror function. This paper went beyond the form of dealing with smart mirrors only stand alone device the provide information to users, and constructed a network in which smart mirrors are connected, and proposed and implemented a CoMirror system that allows users to talk and share information with other smart mirror users. The CoMirror system has a structure in which several CoMirror clients are connected on one CoMirror server. The CoMirror client consists of Raspberry Pi, a mirror film, a touch pad, a display device, an web camera, etc. The server has functions such as face learning and recognition, user management, a relay role for exchanging messages between clients, and setting up for video call. Users can communicate with other CoMirror users via the server, such as text, image, and audio messages, as well as 1:1 video call.

Panorama Background Generation and Object Tracking using Pan-Tilt-Zoom Camera (Pan-Tilt-Zoom 카메라를 이용한 파노라마 배경 생성과 객체 추적)

  • Paek, In-Ho;Im, Jae-Hyun;Park, Kyoung-Ju;Paik, Jun-Ki
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.45 no.3
    • /
    • pp.55-63
    • /
    • 2008
  • This paper presents a panorama background generation and object tracking technique using a Pan-Tilt-Zoom camera. The proposed method estimates local motion vectors rapidly using phase correlation matching at the prespecified multiple local regions, and it makes minimized estimation error by vector quantization. We obtain the required image patches, by estimating the overlapped region using local motion vectors, we can then project the images to cylinder and realign the images to make the panoramic image. The object tracking is performed by extracting object's motion and by separating foreground from input image using background subtraction. The proposed PTZ-based object tracking method can efficiently generated a stable panorama background, which covers up to 360 degree FOV The proposed algorithm is designed for real-time implementation and it can be applied to many commercial applications such as object shape detection and face recognition in various surveillance video systems.

New Scheme for Smoker Detection (흡연자 검출을 위한 새로운 방법)

  • Lee, Jong-seok;Lee, Hyun-jae;Lee, Dong-kyu;Oh, Seoung-jun
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.41 no.9
    • /
    • pp.1120-1131
    • /
    • 2016
  • In this paper, we propose a smoker recognition algorithm, detecting smokers in a video sequence in order to prevent fire accidents. We use description-based method in hierarchical approaches to recognize smoker's activity, the algorithm consists of background subtraction, object detection, event search, event judgement. Background subtraction generates slow-motion and fast-motion foreground image from input image using Gaussian mixture model with two different learning-rate. Then, it extracts object locations in the slow-motion image using chain-rule based contour detection. For each object, face is detected by using Haar-like feature and smoke is detected by reflecting frequency and direction of smoke in fast-motion foreground. Hand movements are detected by motion estimation. The algorithm examines the features in a certain interval and infers that whether the object is a smoker. It robustly can detect a smoker among different objects while achieving real-time performance.

Social Network Analysis of TV Drama via Location Knowledge-learned Deep Hypernetworks (장소 정보를 학습한 딥하이퍼넷 기반 TV드라마 소셜 네트워크 분석)

  • Nan, Chang-Jun;Kim, Kyung-Min;Zhang, Byoung-Tak
    • KIISE Transactions on Computing Practices
    • /
    • v.22 no.11
    • /
    • pp.619-624
    • /
    • 2016
  • Social-aware video displays not only the relationships between characters but also diverse information on topics such as economics, politics and culture as a story unfolds. Particularly, the speaking habits and behavioral patterns of people in different situations are very important for the analysis of social relationships. However, when dealing with this dynamic multi-modal data, it is difficult for a computer to analyze the drama data effectively. To solve this problem, previous studies employed the deep concept hierarchy (DCH) model to automatically construct and analyze social networks in a TV drama. Nevertheless, since location knowledge was not included, they can only analyze the social network as a whole in stories. In this research, we include location knowledge and analyze the social relations in different locations. We adopt data from approximately 4400 minutes of a TV drama Friends as our dataset. We process face recognition on the characters by using a convolutional- recursive neural networks model and utilize a bag of features model to classify scenes. Then, in different scenes, we establish the social network between the characters by using a deep concept hierarchy model and analyze the change in the social network while the stories unfold.

Vision-based Low-cost Walking Spatial Recognition Algorithm for the Safety of Blind People (시각장애인 안전을 위한 영상 기반 저비용 보행 공간 인지 알고리즘)

  • Sunghyun Kang;Sehun Lee;Junho Ahn
    • Journal of Internet Computing and Services
    • /
    • v.24 no.6
    • /
    • pp.81-89
    • /
    • 2023
  • In modern society, blind people face difficulties in navigating common environments such as sidewalks, elevators, and crosswalks. Research has been conducted to alleviate these inconveniences for the visually impaired through the use of visual and audio aids. However, such research often encounters limitations when it comes to practical implementation due to the high cost of wearable devices, high-performance CCTV systems, and voice sensors. In this paper, we propose an artificial intelligence fusion algorithm that utilizes low-cost video sensors integrated into smartphones to help blind people safely navigate their surroundings during walking. The proposed algorithm combines motion capture and object detection algorithms to detect moving people and various obstacles encountered during walking. We employed the MediaPipe library for motion capture to model and detect surrounding pedestrians during motion. Additionally, we used object detection algorithms to model and detect various obstacles that can occur during walking on sidewalks. Through experimentation, we validated the performance of the artificial intelligence fusion algorithm, achieving accuracy of 0.92, precision of 0.91, recall of 0.99, and an F1 score of 0.95. This research can assist blind people in navigating through obstacles such as bollards, shared scooters, and vehicles encountered during walking, thereby enhancing their mobility and safety.

Abnormal Crowd Behavior Detection via H.264 Compression and SVDD in Video Surveillance System (H.264 압축과 SVDD를 이용한 영상 감시 시스템에서의 비정상 집단행동 탐지)

  • Oh, Seung-Geun;Lee, Jong-Uk;Chung, Yongw-Ha;Park, Dai-Hee
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.21 no.6
    • /
    • pp.183-190
    • /
    • 2011
  • In this paper, we propose a prototype system for abnormal sound detection and identification which detects and recognizes the abnormal situations by means of analyzing audio information coming in real time from CCTV cameras under surveillance environment. The proposed system is composed of two layers: The first layer is an one-class support vector machine, i.e., support vector data description (SVDD) that performs rapid detection of abnormal situations and alerts to the manager. The second layer classifies the detected abnormal sound into predefined class such as 'gun', 'scream', 'siren', 'crash', 'bomb' via a sparse representation classifier (SRC) to cope with emergency situations. The proposed system is designed in a hierarchical manner via a mixture of SVDD and SRC, which has desired characteristics as follows: 1) By fast detecting abnormal sound using SVDD trained with only normal sound, it does not perform the unnecessary classification for normal sound. 2) It ensures a reliable system performance via a SRC that has been successfully applied in the field of face recognition. 3) With the intrinsic incremental learning capability of SRC, it can actively adapt itself to the change of a sound database. The experimental results with the qualitative analysis illustrate the efficiency of the proposed method.

Analysis of Space Use Patterns of Public Library Users through AI Cameras (AI 카메라를 활용한 공공도서관 이용자의 공간이용행태 분석 연구)

  • Gyuhwan Kim;Do-Heon Jeong
    • Journal of the Korean Society for Library and Information Science
    • /
    • v.57 no.4
    • /
    • pp.333-351
    • /
    • 2023
  • This study investigates user behavior in library spaces through the lens of AI camera analytics. By leveraging the face recognition and tracking capabilities of AI cameras, we accurately identified the gender and age of visitors and meticulously collected video data to track their movements. Our findings revealed that female users slightly outnumbered male users and the dominant age group was individuals in their 30s. User visits peaked between Tuesday to Friday, with the highest footfall recorded between 14:00 and 15:00 pm, while visits decreased over the weekend. Most visitors utilized one or two specific spaces, frequently consulting the information desk for inquiries, checking out/returning items, or using the rest area for relaxation. The library stacks were used approximately twice as much as they were avoided. The most frequented subject areas were Philosophy(100), Religion(200), Social Sciences(300), Science(400), Technology(500), and Literature(800), with Literature(800) and Religion(200) displaying the most intersections with other areas. By categorizing users into five clusters based on space utilization patterns, we discerned varying objectives and subject interests, providing insights for future library service enhancements. Moreover, the study underscores the need to address the associated costs and privacy concerns when considering the broader application of AI camera analytics in library settings.