• Title/Summary/Keyword: Eye Landmark Detection

Search Result 6, Processing Time 0.022 seconds

Deep Learning-based Gaze Direction Vector Estimation Network Integrated with Eye Landmark Localization (딥 러닝 기반의 눈 랜드마크 위치 검출이 통합된 시선 방향 벡터 추정 네트워크)

  • Joo, Heeyoung;Ko, Min-Soo;Song, Hyok
    • Journal of Broadcast Engineering
    • /
    • v.26 no.6
    • /
    • pp.748-757
    • /
    • 2021
  • In this paper, we propose a gaze estimation network in which eye landmark position detection and gaze direction vector estimation are integrated into one deep learning network. The proposed network uses the Stacked Hourglass Network as a backbone structure and is largely composed of three parts: a landmark detector, a feature map extractor, and a gaze direction estimator. The landmark detector estimates the coordinates of 50 eye landmarks, and the feature map extractor generates a feature map of the eye image for estimating the gaze direction. And the gaze direction estimator estimates the final gaze direction vector by combining each output result. The proposed network was trained using virtual synthetic eye images and landmark coordinate data generated through the UnityEyes dataset, and the MPIIGaze dataset consisting of real human eye images was used for performance evaluation. Through the experiment, the gaze estimation error showed a performance of 3.9, and the estimation speed of the network was 42 FPS (Frames per second).

An Effective Retinal Vessel and Landmark Detection Algorithm in RGB images

  • Jung Eun-Hwa
    • International Journal of Contents
    • /
    • v.2 no.3
    • /
    • pp.27-32
    • /
    • 2006
  • We present an effective algorithm for automatic tracing of retinal vessel structure and vascular landmark extraction of bifurcations and ending points. In this paper we deal with vascular patterns from RGB images for personal identification. Vessel tracing algorithms are of interest in a variety of biometric and medical application such as personal identification, biometrics, and ophthalmic disorders like vessel change detection. However eye surface vasculature tracing in RGB images has many problems which are subject to improper illumination, glare, fade-out, shadow and artifacts arising from reflection, refraction, and dispersion. The proposed algorithm on vascular tracing employs multi-stage processing of ten-layers as followings: Image Acquisition, Image Enhancement by gray scale retinal image enhancement, reducing background artifact and illuminations and removing interlacing minute characteristics of vessels, Vascular Structure Extraction by connecting broken vessels, extracting vascular structure using eight directional information, and extracting retinal vascular structure, and Vascular Landmark Extraction by extracting bifurcations and ending points. The results of automatic retinal vessel extraction using jive different thresholds applied 34 eye images are presented. The results of vasculature tracing algorithm shows that the suggested algorithm can obtain not only robust and accurate vessel tracing but also vascular landmarks according to thresholds.

  • PDF

Improving Eye-gaze Mouse System Using Mouth Open Detection and Pop Up Menu (입 벌림 인식과 팝업 메뉴를 이용한 시선추적 마우스 시스템 성능 개선)

  • Byeon, Ju Yeong;Jung, Keechul
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.12
    • /
    • pp.1454-1463
    • /
    • 2020
  • An important factor in eye-tracking PC interface for general paralyzed patients is the implementation of the mouse interface, for manipulating the GUI. With a successfully implemented mouse interface, users can generate mouse events exactly at the point of their choosing. However, it is difficult to define this interaction in the eye-tracking interface. This problem has been defined as the Midas touch problem and has been a major focus of eye-tracking research. There have been many attempts to solve this problem using blink, voice input, etc. However, it was not suitable for general paralyzed patients because some of them cannot wink or speak. In this paper, we propose a mouth-pop-up, eye-tracking mouse interface that solves the Midas touch problem as well as becoming a suitable interface for general paralyzed patients using a common RGB camera. The interface presented in this paper implements a mouse interface that detects the opening and closing of the mouth to activate a pop-up menu that the user can select the mouse event. After implementation, a performance experiment was conducted. As a result, we found that the number of malfunctions and the time to perform tasks were reduced compared to the existing method.

Applying MetaHuman Facial Animation with MediaPipe: An Alternative Solution to Live Link iPhone.

  • Balgum Song;Arminas Baronas
    • International journal of advanced smart convergence
    • /
    • v.13 no.3
    • /
    • pp.191-198
    • /
    • 2024
  • This paper presents an alternative solution for applying MetaHuman facial animations using MediaPipe, providing a versatile option to the Live Link iPhone system. Our approach involves capturing facial expressions with various camera devices, including webcams, laptop cameras, and Android phones, processing the data for landmark detection, and applying these landmarks in Unreal Engine Blueprint to animate MetaHuman characters in real-time. Techniques such as the Eye Aspect Ratio (EAR) for blink detection and the One Euro Filter for data smoothing ensure accurate and responsive animations. Experimental results demonstrate that our system provides a cost-effective and flexible alternative for iPhone non-users, enhancing the accessibility of advanced facial capture technology for applications in digital media and interactive environments. This research offers a practical and adaptable method for real-time facial animation, with future improvements aimed at integrating more sophisticated emotion detection features.

A Driver's Condition Warning System using Eye Aspect Ratio (눈 영상비를 이용한 운전자 상태 경고 시스템)

  • Shin, Moon-Chang;Lee, Won-Young
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.15 no.2
    • /
    • pp.349-356
    • /
    • 2020
  • This paper introduces the implementation of a driver's condition warning system using eye aspect ratio to prevent a car accident. The proposed driver's condition warning system using eye aspect ratio consists of a camera, that is required to detect eyes, the Raspberrypie that processes information on eyes from the camera, buzzer and vibrator, that are required to warn the driver. In order to detect and recognize driver's eyes, the histogram of oriented gradients and face landmark estimation based on deep-learning are used. Initially the system calculates the eye aspect ratio of the driver from 6 coordinates around the eye and then gets each eye aspect ratio values when the eyes are opened and closed. These two different eye aspect ratio values are used to calculate the threshold value that is necessary to determine the eye state. Because the threshold value is adaptively determined according to the driver's eye aspect ratio, the system can use the optimal threshold value to determine the driver's condition. In addition, the system synthesizes an input image from the gray-scaled and LAB model images to operate in low lighting conditions.

Non-contact Input Method based on Face Recognition and Pyautogui Mouse Control (얼굴 인식과 Pyautogui 마우스 제어 기반의 비접촉식 입력 기법)

  • Park, Sung-jin;Shin, Ye-eun;Lee, Byung-joon;Oh, Ha-young
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.9
    • /
    • pp.1279-1292
    • /
    • 2022
  • This study proposes a non-contact input method based on face recognition and Pyautogui mouse control as a system that can help users who have difficulty using input devices such as conventional mouse due to physical discomfort. This study includes features that help web surfing more conveniently, especially screen zoom, scroll function, and also solves the problem of eye fatigue, which has been suggested as a limitation in existing non-contact input systems. In addition, various set values can be adjusted in consideration of individual physical differences and Internet usage habits. Furthermore, no high-performance CPU or GPU environment is required, and no separate tracker devices or high-performance cameras are required. Through these studies, we intended to contribute to the realization of barrier-free access by increasing the web accessibility of the disabled and the elderly who find it difficult to use web content.