• Title/Summary/Keyword: Eye-gaze Interface

Search Result 36, Processing Time 0.026 seconds

A Study on the Korean Characters(Hangul) Input Method of Using Eye-gaze Interface (Eye-gaze Interface를 이용한 한글입력시스템 한글입력시스템 개발에 관한 연구)

  • 서한석;김종기;김치용
    • Proceedings of the Korea Multimedia Society Conference
    • /
    • 2004.05a
    • /
    • pp.533-537
    • /
    • 2004
  • 고도 정보사회의 진전에 있어 정보의 수집, 정보의 전달 등을 위해, 여러 IT기기가 사용 되어지고 있다. 이러한 상황 속에서 IT기기를 장애자나 고령자를 포함, 누구나가 쉽게 이용 할 수 있도록 하는 유니버설디자인은 매우 중요한 일이다. 특히 상체부자유자의 인터페이스 장치로서 시선입력장치가 주목 받고 있다. 이러한 시대적인 흐름에서 본 연구에서는 Eye-gaze Interface를 이용한 시선입력만으로 컴퓨터상에서 한글입력이 가능한 한글 입력 시스템을 개발 제안한다. 또한 본 연구의 목적으로는 기존의 키보드, 마우스에 의한 입력장치와 Eye-gaze Interface를 이용한 입력 장치와의 정점과 단점을 비교 분석하고, 정보 사회에 있어서의 장애자와 고령자들의 Barrier Free에 관한 연구를 기본 목적으로 하고 있다 본 연구의 Eye-gaze Interface는 Eye Tech Digital System회사의 Quick Glance System를 사용한다 장애자와 고령자를 대상으로 한 사용자 평가 실험에 의해 실험을 진행 하였으며, 한글의 음운규칙에 관한 연구를 기초로 한글뿐만 아니라 영어, 일본어가 입력 가능한 입력 시스템을 제안한다. 이것은 컴퓨터 조작이 곤란한 장애자와 고령자가 일반인과의 자연스러운 커뮤니케이션이 가능하도록 한 것이다. 향후 Eye-gaze Interface를 이용한 응용분야 또한 게임, 영상, Eye Book, 자동차운전 둥 다양한 분야에 적용 가능하리라 기대된다.

  • PDF

A Study on the Hangul Input Methodology for Eye-gaze Interface (시선 입력 장치에 의한 한글 입력 시스템 설계에 관한 연구)

  • Seo Han-Sok;Kim Chee-Yong
    • Journal of Digital Contents Society
    • /
    • v.5 no.3
    • /
    • pp.239-244
    • /
    • 2004
  • New developments in IT already impact wide segments of a young and mobile population. It is evident that applications of information technology can be of equal benefit to the aged and the disabled. `Eye-Gaze'(EGI) technology was designed for people with paralysis in the upper body. There is a compeling need for a dedicated Korean Language interface for this system. The purpose of this study is to research 'Barrier Free' software using a control group of the mobility impaired to assess the Eye-Gaze Interface in the context of more conventional input methods. TheEGI of this study uses Quick Glance System of Eye Tech Digital Systems. The study will be evaluated on criteria based upon the needs of those with specific disabilities and mobility problems associated with aging. We also intend to explore applications of the Eye-Gaze Interface for English and Japanese devises, based upon our study using the Hangul phonology.

  • PDF

A New Eye Tracking Method as a Smartphone Interface

  • Lee, Eui Chul;Park, Min Woo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.7 no.4
    • /
    • pp.834-848
    • /
    • 2013
  • To effectively use these functions many kinds of human-phone interface are used such as touch, voice, and gesture. However, the most important touch interface cannot be used in case of hand disabled person or busy both hands. Although eye tracking is a superb human-computer interface method, it has not been applied to smartphones because of the small screen size, the frequently changing geometric position between the user's face and phone screen, and the low resolution of the frontal cameras. In this paper, a new eye tracking method is proposed to act as a smartphone user interface. To maximize eye image resolution, a zoom lens and three infrared LEDs are adopted. Our proposed method has following novelties. Firstly, appropriate camera specification and image resolution are analyzed in order to smartphone based gaze tracking method. Secondly, facial movement is allowable in case of one eye region is included in image. Thirdly, the proposed method can be operated in case of both landscape and portrait screen modes. Fourthly, only two LED reflective positions are used in order to calculate gaze position on the basis of 2D geometric relation between reflective rectangle and screen. Fifthly, a prototype mock-up design module is made in order to confirm feasibility for applying to actual smart-phone. Experimental results showed that the gaze estimation error was about 31 pixels at a screen resolution of $480{\times}800$ and the average hit ratio of a $5{\times}4$ icon grid was 94.6%.

A Human-Robot Interface Using Eye-Gaze Tracking System for People with Motor Disabilities

  • Kim, Do-Hyoung;Kim, Jae-Hean;Yoo, Dong-Hyun;Lee, Young-Jin;Chung, Myung-Jin
    • Transactions on Control, Automation and Systems Engineering
    • /
    • v.3 no.4
    • /
    • pp.229-235
    • /
    • 2001
  • Recently, service area has been emerging field f robotic applications. Even though assistant robots play an important role for the disabled and the elderly, they still suffer from operating the robots using conventional interface devices such as joysticks or keyboards. In this paper we propose an efficient computer interface using real-time eye-gaze tracking system. The inputs to the proposed system are images taken by a camera and data from a magnetic sensor. The measured data is sufficient to describe the eye and head movement because the camera and the receiver of a magnetic sensor are stationary with respect to the head. So the proposed system can obtain the eye-gaze direction in spite of head movement as long as the distance between the system and the transmitter of a magnetic position sensor is within 2m. Experimental results show the validity of the proposed system in practical aspect and also verify the feasibility of the system as a new computer interface for the disabled.

  • PDF

Improving Eye-gaze Mouse System Using Mouth Open Detection and Pop Up Menu (입 벌림 인식과 팝업 메뉴를 이용한 시선추적 마우스 시스템 성능 개선)

  • Byeon, Ju Yeong;Jung, Keechul
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.12
    • /
    • pp.1454-1463
    • /
    • 2020
  • An important factor in eye-tracking PC interface for general paralyzed patients is the implementation of the mouse interface, for manipulating the GUI. With a successfully implemented mouse interface, users can generate mouse events exactly at the point of their choosing. However, it is difficult to define this interaction in the eye-tracking interface. This problem has been defined as the Midas touch problem and has been a major focus of eye-tracking research. There have been many attempts to solve this problem using blink, voice input, etc. However, it was not suitable for general paralyzed patients because some of them cannot wink or speak. In this paper, we propose a mouth-pop-up, eye-tracking mouse interface that solves the Midas touch problem as well as becoming a suitable interface for general paralyzed patients using a common RGB camera. The interface presented in this paper implements a mouse interface that detects the opening and closing of the mouth to activate a pop-up menu that the user can select the mouse event. After implementation, a performance experiment was conducted. As a result, we found that the number of malfunctions and the time to perform tasks were reduced compared to the existing method.

Object Magnification and Voice Command in Gaze Interface for the Upper Limb Disabled (상지장애인을 위한 시선 인터페이스에서의 객체 확대 및 음성 명령 인터페이스 개발)

  • Park, Joo Hyun;Jo, Se-Ran;Lim, Soon-Bum
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.7
    • /
    • pp.903-912
    • /
    • 2021
  • Eye tracking research for upper limb disabilities is showing an effect in the aspect of device control. However, the reality is that it is not enough to perform web interaction with only eye tracking technology. In the Eye-Voice interface, a previous study, in order to solve the problem that the existing gaze tracking interfaces cause a malfunction of pointer execution, a gaze tracking interface supplemented with a voice command was proposed. In addition, the reduction of the malfunction rate of the pointer was confirmed through a comparison experiment with the existing interface. In this process, the difficulty of pointing due to the small size of the execution object in the web environment was identified as another important problem of malfunction. In this study, we propose an auto-magnification interface of objects so that people with upper extremities can freely click web contents by improving the problem that it was difficult to point and execute due to the high density of execution objects and their arrangements in web pages.

A New Ergonomic Interface System for the Disabled Person (장애인을 위한 새로운 감성 인터페이스 연구)

  • Heo, Hwan;Lee, Ji-Woo;Lee, Won-Oh;Lee, Eui-Chul;Park, Kang-Ryoung
    • Journal of the Ergonomics Society of Korea
    • /
    • v.30 no.1
    • /
    • pp.229-235
    • /
    • 2011
  • Objective: Making a new ergonomic interface system based on camera vision system, which helps the handicapped in home environment. Background: Enabling the handicapped to manipulate the consumer electronics by the proposed interface system. Method: A wearable device for capturing the eye image using a near-infrared(NIR) camera and illuminators is proposed for tracking eye gaze position(Heo et al., 2011). A frontal viewing camera is attached to the wearable device, which can recognize the consumer electronics to be controlled(Heo et al., 2011). And the amount of user's eye fatigue can be measured based on eye blink rate, and in case that the user's fatigue exceeds in the predetermined level, the proposed system can automatically change the mode of gaze based interface into that of manual selection. Results: The experimental results showed that the gaze estimation error of the proposed method was 1.98 degrees with the successful recognition of the object by the frontal viewing camera(Heo et al., 2011). Conclusion: We made a new ergonomic interface system based on gaze tracking and object recognition Application: The proposed system can be used for helping the handicapped in home environment.

Design of Korean eye-typing interfaces based on multilevel input system (단계식 입력 체계를 이용한 시선 추적 기반의 한글 입력 인터페이스 설계)

  • Kim, Hojoong;Woo, Sung-kyung;Lee, Kunwoo
    • Journal of the HCI Society of Korea
    • /
    • v.12 no.4
    • /
    • pp.37-44
    • /
    • 2017
  • Eye-typing is one kind of human-computer interactive input system which is implemented by location data of gaze. It is widely used as an input system for paralytics because it does not require physical motions other than the eye movement. However, eye-typing interface based on Korean character has not been suggested yet. Thus, this research aims to implement the eye-typing interface optimized for Korean. To begin with, design objectives were established based on the features of eye-typing: significant noise and Midas touch problem. Multilevel input system was introduced to deal with noise, and an area free from input button was applied to solve Midas touch problem. Then, two types of eye-typing interfaces were suggested on phonological consideration of Korean where each syllable is generated from combination of several phonemes. Named as consonant-vowel integrated interface and separated interface, the two interfaces are designed to input Korean in phases through grouped phonemes. Finally, evaluation procedures composed of comparative experiments against the conventional Double-Korean keyboard interface, and analysis on flow of gaze were conducted. As a result, newly designed interfaces showed potential to be applied as practical tools for eye-typing.