• Title/Summary/Keyword: face detect

Search Result 380, Processing Time 0.031 seconds

Robust 3D Facial Landmark Detection Using Angular Partitioned Spin Images (각 분할 스핀 영상을 사용한 3차원 얼굴 특징점 검출 방법)

  • Kim, Dong-Hyun;Choi, Kang-Sun
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.5
    • /
    • pp.199-207
    • /
    • 2013
  • Spin images representing efficiently surface features of 3D mesh models have been used to detect facial landmark points. However, at a certain point, different normal direction can lead to quite different spin images. Moreover, since 3D points are projected to the 2D (${\alpha}-{\beta}$) space during spin image generation, surface features cannot be described clearly. In this paper, we present a method to detect 3D facial landmark using improved spin images by partitioning the search area with respect to angle. By generating sub-spin images for angular partitioned 3D spaces, more unique features describing corresponding surfaces can be obtained, and improve the performance of landmark detection. In order to generate spin images robust to inaccurate surface normal direction, we utilize on averaging surface normal with its neighboring normal vectors. The experimental results show that the proposed method increases the accuracy in landmark detection by about 34% over a conventional method.

A Driver's Condition Warning System using Eye Aspect Ratio (눈 영상비를 이용한 운전자 상태 경고 시스템)

  • Shin, Moon-Chang;Lee, Won-Young
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.15 no.2
    • /
    • pp.349-356
    • /
    • 2020
  • This paper introduces the implementation of a driver's condition warning system using eye aspect ratio to prevent a car accident. The proposed driver's condition warning system using eye aspect ratio consists of a camera, that is required to detect eyes, the Raspberrypie that processes information on eyes from the camera, buzzer and vibrator, that are required to warn the driver. In order to detect and recognize driver's eyes, the histogram of oriented gradients and face landmark estimation based on deep-learning are used. Initially the system calculates the eye aspect ratio of the driver from 6 coordinates around the eye and then gets each eye aspect ratio values when the eyes are opened and closed. These two different eye aspect ratio values are used to calculate the threshold value that is necessary to determine the eye state. Because the threshold value is adaptively determined according to the driver's eye aspect ratio, the system can use the optimal threshold value to determine the driver's condition. In addition, the system synthesizes an input image from the gray-scaled and LAB model images to operate in low lighting conditions.

Annotation Method based on Face Area for Efficient Interactive Video Authoring (효과적인 인터랙티브 비디오 저작을 위한 얼굴영역 기반의 어노테이션 방법)

  • Yoon, Ui Nyoung;Ga, Myeong Hyeon;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.1
    • /
    • pp.83-98
    • /
    • 2015
  • Many TV viewers use mainly portal sites in order to retrieve information related to broadcast while watching TV. However retrieving information that people wanted needs a lot of time to retrieve the information because current internet presents too much information which is not required. Consequentially, this process can't satisfy users who want to consume information immediately. Interactive video is being actively investigated to solve this problem. An interactive video provides clickable objects, areas or hotspots to interact with users. When users click object on the interactive video, they can see additional information, related to video, instantly. The following shows the three basic procedures to make an interactive video using interactive video authoring tool: (1) Create an augmented object; (2) Set an object's area and time to be displayed on the video; (3) Set an interactive action which is related to pages or hyperlink; However users who use existing authoring tools such as Popcorn Maker and Zentrick spend a lot of time in step (2). If users use wireWAX then they can save sufficient time to set object's location and time to be displayed because wireWAX uses vision based annotation method. But they need to wait for time to detect and track object. Therefore, it is required to reduce the process time in step (2) using benefits of manual annotation method and vision-based annotation method effectively. This paper proposes a novel annotation method allows annotator to easily annotate based on face area. For proposing new annotation method, this paper presents two steps: pre-processing step and annotation step. The pre-processing is necessary because system detects shots for users who want to find contents of video easily. Pre-processing step is as follow: 1) Extract shots using color histogram based shot boundary detection method from frames of video; 2) Make shot clusters using similarities of shots and aligns as shot sequences; and 3) Detect and track faces from all shots of shot sequence metadata and save into the shot sequence metadata with each shot. After pre-processing, user can annotates object as follow: 1) Annotator selects a shot sequence, and then selects keyframe of shot in the shot sequence; 2) Annotator annotates objects on the relative position of the actor's face on the selected keyframe. Then same objects will be annotated automatically until the end of shot sequence which has detected face area; and 3) User assigns additional information to the annotated object. In addition, this paper designs the feedback model in order to compensate the defects which are wrong aligned shots, wrong detected faces problem and inaccurate location problem might occur after object annotation. Furthermore, users can use interpolation method to interpolate position of objects which is deleted by feedback. After feedback user can save annotated object data to the interactive object metadata. Finally, this paper shows interactive video authoring system implemented for verifying performance of proposed annotation method which uses presented models. In the experiment presents analysis of object annotation time, and user evaluation. First, result of object annotation average time shows our proposed tool is 2 times faster than existing authoring tools for object annotation. Sometimes, annotation time of proposed tool took longer than existing authoring tools, because wrong shots are detected in the pre-processing. The usefulness and convenience of the system were measured through the user evaluation which was aimed at users who have experienced in interactive video authoring system. Recruited 19 experts evaluates of 11 questions which is out of CSUQ(Computer System Usability Questionnaire). CSUQ is designed by IBM for evaluating system. Through the user evaluation, showed that proposed tool is useful for authoring interactive video than about 10% of the other interactive video authoring systems.

Real-Time Head Tracking using Adaptive Boosting in Surveillance (서베일런스에서 Adaptive Boosting을 이용한 실시간 헤드 트래킹)

  • Kang, Sung-Kwan;Lee, Jung-Hyun
    • Journal of Digital Convergence
    • /
    • v.11 no.2
    • /
    • pp.243-248
    • /
    • 2013
  • This paper proposes an effective method using Adaptive Boosting to track a person's head in complex background. By only one way to feature extraction methods are not sufficient for modeling a person's head. Therefore, the method proposed in this paper, several feature extraction methods for the accuracy of the detection head running at the same time. Feature Extraction for the imaging of the head was extracted using sub-region and Haar wavelet transform. Sub-region represents the local characteristics of the head, Haar wavelet transform can indicate the frequency characteristics of face. Therefore, if we use them to extract the features of face, effective modeling is possible. In the proposed method to track down the man's head from the input video in real time, we ues the results after learning Harr-wavelet characteristics of the three types using AdaBoosting algorithm. Originally the AdaBoosting algorithm, there is a very long learning time, if learning data was changes, and then it is need to be performed learning again. In order to overcome this shortcoming, in this research propose efficient method using cascade AdaBoosting. This method reduces the learning time for the imaging of the head, and can respond effectively to changes in the learning data. The proposed method generated classifier with excellent performance using less learning time and learning data. In addition, this method accurately detect and track head of person from a variety of head data in real-time video images.

CT Number Measurement of Residual Foreign Bodies in Face (안면부에 잔류된 다양한 이물질을 측정한 CT 계수)

  • Wee, Syeo Young;Choi, Hwan Jun;Kim, Mi Sun;Choi, Chang Yong
    • Archives of Plastic Surgery
    • /
    • v.35 no.4
    • /
    • pp.423-430
    • /
    • 2008
  • Purpose: Computed tomography theoretically should improve detection of foreign bodies and provide more information of adjacent soft tissues. And the CT scanner and PACS program proved to be an excellent instrument for detection and localization of most facial foreign bodies above certain minimum levels of detectability. The severity of injury in penetrating trauma to the face, it is often underestimated by physical examination. Diagnosis of a retained foreign object is always critical. Methods: From March, 2005 to February 2008 a study was done with 200 patients who had facial trauma. Axial and coronal CT images were obtained with a General Electric(Milwaukee, Wis) 9800 CT scanner at 130 kV, 90 mA, with a 2-mm section thickness and a $512{\times}512$ matrix. Results: Axial and coronal CT images at various window widths should be used as the first imaging modality to detect facial foreign bodies. The attenuation coefficients for the metallic and nonmetallic foreign bodies ranged from -437 to +3071 HU. As a general rule, metallic foreign bodies produced more Hounsfield artifacts than nonmetallic foreign bodies, thus providing a clue to their composition. All of the metallic foreign bodies were represented by a single peak and had a maximum attenuation coefficient of +3071 HU. Of the nonmetallic foreign bodies, glass had an attenuation coefficient that ranged from +105 to +2039, while plastic had a much lower coefficient that ranged from -62 to -35. wood had the lowest range of attenuation coefficients: -491 to -437. Conclusion: The PACS program allows one to distinguish metallic from nonmetallic foreign bodies and to individually identify the specific composition of many nonmetallic foreign bodies. This program does not, however, allow identification of the specific composition of a metallic foreign body. We recommend this type of software program for CT scanning of any patient with an injury to the face in which a foreign body is suspected.

3D Facial Animation with Head Motion Estimation and Facial Expression Cloning (얼굴 모션 추정과 표정 복제에 의한 3차원 얼굴 애니메이션)

  • Kwon, Oh-Ryun;Chun, Jun-Chul
    • The KIPS Transactions:PartB
    • /
    • v.14B no.4
    • /
    • pp.311-320
    • /
    • 2007
  • This paper presents vision-based 3D facial expression animation technique and system which provide the robust 3D head pose estimation and real-time facial expression control. Many researches of 3D face animation have been done for the facial expression control itself rather than focusing on 3D head motion tracking. However, the head motion tracking is one of critical issues to be solved for developing realistic facial animation. In this research, we developed an integrated animation system that includes 3D head motion tracking and facial expression control at the same time. The proposed system consists of three major phases: face detection, 3D head motion tracking, and facial expression control. For face detection, with the non-parametric HT skin color model and template matching, we can detect the facial region efficiently from video frame. For 3D head motion tracking, we exploit the cylindrical head model that is projected to the initial head motion template. Given an initial reference template of the face image and the corresponding head motion, the cylindrical head model is created and the foil head motion is traced based on the optical flow method. For the facial expression cloning we utilize the feature-based method, The major facial feature points are detected by the geometry of information of the face with template matching and traced by optical flow. Since the locations of varying feature points are composed of head motion and facial expression information, the animation parameters which describe the variation of the facial features are acquired from geometrically transformed frontal head pose image. Finally, the facial expression cloning is done by two fitting process. The control points of the 3D model are varied applying the animation parameters to the face model, and the non-feature points around the control points are changed by use of Radial Basis Function(RBF). From the experiment, we can prove that the developed vision-based animation system can create realistic facial animation with robust head pose estimation and facial variation from input video image.

Vision-based Low-cost Walking Spatial Recognition Algorithm for the Safety of Blind People (시각장애인 안전을 위한 영상 기반 저비용 보행 공간 인지 알고리즘)

  • Sunghyun Kang;Sehun Lee;Junho Ahn
    • Journal of Internet Computing and Services
    • /
    • v.24 no.6
    • /
    • pp.81-89
    • /
    • 2023
  • In modern society, blind people face difficulties in navigating common environments such as sidewalks, elevators, and crosswalks. Research has been conducted to alleviate these inconveniences for the visually impaired through the use of visual and audio aids. However, such research often encounters limitations when it comes to practical implementation due to the high cost of wearable devices, high-performance CCTV systems, and voice sensors. In this paper, we propose an artificial intelligence fusion algorithm that utilizes low-cost video sensors integrated into smartphones to help blind people safely navigate their surroundings during walking. The proposed algorithm combines motion capture and object detection algorithms to detect moving people and various obstacles encountered during walking. We employed the MediaPipe library for motion capture to model and detect surrounding pedestrians during motion. Additionally, we used object detection algorithms to model and detect various obstacles that can occur during walking on sidewalks. Through experimentation, we validated the performance of the artificial intelligence fusion algorithm, achieving accuracy of 0.92, precision of 0.91, recall of 0.99, and an F1 score of 0.95. This research can assist blind people in navigating through obstacles such as bollards, shared scooters, and vehicles encountered during walking, thereby enhancing their mobility and safety.

A Research on Object Detection Technology for the Visually Impaired (시각장애인을 위한 사물 감지 기술 연구)

  • Jeong, Yeon-Kyu;Kim, Byung-Gyu;Lee, Jeong-Bae
    • The KIPS Transactions:PartB
    • /
    • v.19B no.4
    • /
    • pp.225-230
    • /
    • 2012
  • In this paper, a blind person using a white cane as an adjunct of the things available sensing technology has been implemented. Sensing technology to implement things ultrasonic sensors and a webcam was used to process the data from the server computer. Ultrasonic sensors detect objects within 4meter people distinguish between those things that if the results based on the results will sound off. In this study, ultrasonic sensors, object recognition and human perception with the introduction of techniques and technologies developed for detecting objects in the lives of the visually impaired is expected to be greater usability.

Analytical fault tolerant navigation system for an aerospace launch vehicle using sliding mode observer

  • Hasani, Mahdi;Roshanian, Jafar;Khoshnooda, A. Majid
    • Advances in aircraft and spacecraft science
    • /
    • v.4 no.1
    • /
    • pp.53-64
    • /
    • 2017
  • Aerospace Launch Vehicles (ALV) are generally designed with high reliability to operate in complete security through fault avoidance practices. However, in spite of such precaution, fault occurring is inevitable. Hence, there is a requirement for on-board fault recovery without significant degradation in the ALV performance. The present study develops an advanced fault recovery strategy to improve the reliability of an Aerospace Launch Vehicle (ALV) navigation system. The proposed strategy contains fault detection features and can reconfigure the system against common faults in the ALV navigation system. For this purpose, fault recovery system is constructed to detect and reconfigure normal navigation faults based on the sliding mode observer (SMO) theory. In the face of pitch channel sensor failure, the original gyro faults are reconstructed using SMO theory and by correcting the faulty measurement, the pitch-rate gyroscope output is constructed to provide fault tolerant navigation solution. The novel aspect of the paper is employing SMO as an online tuning of analytical fault recovery solution against unforeseen variations due to its hardware/software property. In this regard, a nonlinear model of the ALV is simulated using specific navigation failures and the results verified the feasibility of the proposed system. Simulation results and sensitivity analysis show that the proposed techniques can produce more effective estimation results than those of the previous techniques, against sensor failures.

Detection And Countermeasure Scheme For Call-Disruption Attacks On SIP-Based Voip Services

  • Ryu, Jea-Tek;Roh, Byeong-Hee;Ryu, Ki-Yeol;Yoon, Myung-Chul
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.6 no.7
    • /
    • pp.1854-1873
    • /
    • 2012
  • Owing to its simplicity and flexibility, the session initiation protocol (SIP) has been widely adopted as a major session-management protocol for Internet telephony or Voice-over IP (VoIP) services. However, SIP has faced various types of security threats. Call-disruption attacks are some of the most severe threats they face, and can greatly inconvenience consumers. In this paper, we analyze such SIP call-disruption attacks, and propose a method for detecting and counteracting them by extending the SIP INFO method with authentication. Using the proposed method, both the target user and the SIP server can detect the existence of a call-disruption attack on a user and counteract the attack. We demonstrate the effectiveness of the proposed method from the viewpoint of computational complexity by configuring a test-bed with an Asterisk SIP proxy server and an SIP performance (SIPp) emulator.