• Title/Summary/Keyword: hand gesture

Search Result 402, Processing Time 0.032 seconds

Face Detection-based Hand Gesture Recognition in Color and Depth Images (색상 및 거리 영상에서의 얼굴검출 기반 손 제스처 인식)

  • Jeon, Hun-Ki;Ko, Jaepil
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2012.11a
    • /
    • pp.580-582
    • /
    • 2012
  • 본 논문에서는 얼굴검출을 통한 실시간 피부색 모델링과 거리정보를 결합하여 손 영역을 검출하고 손 움직임에 따른 방향 및 원 제스처 인식을 위한 규칙 기반 인식방법을 제안한다. 기존과는 달리 손좌표를 사용하는 대신 기존 프레임과 현재 프레임에서의 손 좌표 차이를 이용하여 제스처 구간을 설정하고 자연스러운 제스처 동작에서의 속도변화를 고려할 수 있도록 한다. 실험 데이터는 5명을 대상으로 4방향과 원을 포함하여 총 5가지 제스처를 10회씩 실행하여 획득하였다. 이들 데이터에 대한 인식 실험에서 97%의 인식률을 보였다.

Dialing Interface Design for Safe Driving using Hand Gesture (손동작을 이용한 운전 안전성을 높이기 위한 전화 다이얼 인터페이스 설계)

  • Jang, WonAng;Lee, DoHoon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2012.04a
    • /
    • pp.449-452
    • /
    • 2012
  • 운전 중에 주의를 분산시키는 요소는 대부분 인터페이스 조작에 있으며 교통사고의 직접적인 원인이 된다. 스마트 자동차에 대한 관심이 높아지면서 운전자 안전에 대한 다양한 연구가 모색되고 있다. 순간의 시선이동으로 인해 판단력과 조작능력을 상실 할 수 있는 현재의 인터페이스는 안전성이 보장 되지 못한다. 본 논문에서는 이러한 운전자의 주의를 분산시키는 요소로 부터 안전성을 확보하기 위해서 차량 내 카메라를 이용하여 손동작을 인식하여 직관적인 제스처로 전화번호를 입력하거나 검색할 수 있는 안전한 인터페이스를 제안한다. 제안한 시스템은 직관적 동작과 TTS(Text To Speech)를 활용하여 사용자 편의성과 안전성을 높였다.

Robot Control using Vision based Hand Gesture Recognition (비전기반 손 제스처 인식을 통한 로봇 컨트롤)

  • Kim, Dae-Soo;Kang, Hang-Bong
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2007.11a
    • /
    • pp.197-200
    • /
    • 2007
  • 본 논문에서는 로봇 컨트롤 시스템을 위해 입력 받은 영상부터 몇 가지의 손 제스처를 인식하는 비전기반 손 제스처 인식방법을 제안한다. 로봇으로부터 입력 받은 이미지는 로봇의 위치, 주변환경, 조명 등 여러 요인에 따라 다양하게 존재한다. 본 논문은 다양한 환경에서 입력되는 영상으로부터 시스템이 로봇 컨트롤을 위해 미리 지정한 몇 가지 제스처를 인식하도록 한다. 먼저 이미지 조명 변화에 강한 손 제스처 인식을 위하여 레티넥스 이미지 정규화를 적용한 후, YCrCb 공간 상에서 입력된 영상에서 손 영역을 검출 후 위치를 추정한다. 인식된 손 영역에서 특징벡터를 추출함으로서 입력 영상내의 존재할 수 있는 손의 크기나 손의 회전각도 등에 상관없이 필요로 하는 제스처를 인식하도록 한다. 제안된 제스처 인식 결과는 로봇컨트롤을 위한 기존의 제스처인식과 비교하여 성능을 측정하였다.

  • PDF

Virtual Object Generation Technique Using Multimodal Interface With Speech and Hand Gesture (음성 및 손동작 결합 인터페이스를 통한 가상객체의 생성)

  • Kim, Changseob;Nam, Hyeongil;Park, Jong-Il
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2019.06a
    • /
    • pp.147-149
    • /
    • 2019
  • 가상현실 기술의 발전으로 보다 많은 사람이 가상현실 콘텐츠를 즐길 수 있게 되었다. PC나 스마트폰과 같은 이전의 콘텐츠 플랫폼과 달리 가상현실에서는 3차원 정보를 전달할 수 있는 인터페이스가 요구된다. 2차원에서 3차원으로의 변화는 보다 높은 자유도를 가지는 반면, 사용자는 새로운 인터페이스에 적응해야 하는 불편함 또한 존재한다. 이러한 불편함을 해소하기 위하여 본 논문에서는 가상현실상에서 음성과 손동작을 결합한 인터페이스를 제안한다. 제안하는 인터페이스는 음성과 손동작은 현실 세계에서의 의사소통을 모방하여 구현하였다. 현실 세계의 의사소통을 모방하였기 때문에 사용자는 추가적인 학습이 없이 가상현실 플랫폼에 보다 쉽게 적응할 수 있다. 또한, 본 논문에서는 가상객체를 생성하는 예제를 통하여 기존의 3차원 입력장치를 대신할 수 있음을 보인다.

  • PDF

Development of Hand-Controlled Transportation Robot (손동작으로 제어 가능한 운송 로봇 개발)

  • Lee, In-kyu;Cho, Young-jun;Kang, Jeong-seok;Lee, Yun-jae;Yoo, Hongseok
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2022.07a
    • /
    • pp.481-482
    • /
    • 2022
  • 본 논문에서는 손동작으로 제어 가능한 운송 로봇을 제안한다. 제안한 시스템에서 로봇은 MediaPipe를 이용하여 실시간으로 사람의 손동작을 인식한다. 또한, 동시적 위치 추적 지도 작성 기법인 SLAM(Simultaneous Localization and Mapping) 기술을 이용하여 로봇이 실내 공간에서 길을 찾고 자율적으로 이동할 수 있게 한다. 개발된 로봇실험을 통하여 로봇이 실시간으로 손동작을 인식하고 동작을 제어하는 것을 확인하였다. 또한, 사전에 작성된 지도를 바탕으로 실내에서 로봇이 자율주행을 하는 것을 확인하였다.

  • PDF

NUI/NUX of the Virtual Monitor Concept using the Concentration Indicator and the User's Physical Features (사용자의 신체적 특징과 뇌파 집중 지수를 이용한 가상 모니터 개념의 NUI/NUX)

  • Jeon, Chang-hyun;Ahn, So-young;Shin, Dong-il;Shin, Dong-kyoo
    • Journal of Internet Computing and Services
    • /
    • v.16 no.6
    • /
    • pp.11-21
    • /
    • 2015
  • As growing interest in Human-Computer Interaction(HCI), research on HCI has been actively conducted. Also with that, research on Natural User Interface/Natural User eXperience(NUI/NUX) that uses user's gesture and voice has been actively conducted. In case of NUI/NUX, it needs recognition algorithm such as gesture recognition or voice recognition. However these recognition algorithms have weakness because their implementation is complex and a lot of time are needed in training because they have to go through steps including preprocessing, normalization, feature extraction. Recently, Kinect is launched by Microsoft as NUI/NUX development tool which attracts people's attention, and studies using Kinect has been conducted. The authors of this paper implemented hand-mouse interface with outstanding intuitiveness using the physical features of a user in a previous study. However, there are weaknesses such as unnatural movement of mouse and low accuracy of mouse functions. In this study, we designed and implemented a hand mouse interface which introduce a new concept called 'Virtual monitor' extracting user's physical features through Kinect in real-time. Virtual monitor means virtual space that can be controlled by hand mouse. It is possible that the coordinate on virtual monitor is accurately mapped onto the coordinate on real monitor. Hand-mouse interface based on virtual monitor concept maintains outstanding intuitiveness that is strength of the previous study and enhance accuracy of mouse functions. Further, we increased accuracy of the interface by recognizing user's unnecessary actions using his concentration indicator from his encephalogram(EEG) data. In order to evaluate intuitiveness and accuracy of the interface, we experimented it for 50 people from 10s to 50s. As the result of intuitiveness experiment, 84% of subjects learned how to use it within 1 minute. Also, as the result of accuracy experiment, accuracy of mouse functions (drag(80.4%), click(80%), double-click(76.7%)) is shown. The intuitiveness and accuracy of the proposed hand-mouse interface is checked through experiment, this is expected to be a good example of the interface for controlling the system by hand in the future.

AdaBoost-based Gesture Recognition Using Time Interval Window Applied Global and Local Feature Vectors with Mono Camera (모노 카메라 영상기반 시간 간격 윈도우를 이용한 광역 및 지역 특징 벡터 적용 AdaBoost기반 제스처 인식)

  • Hwang, Seung-Jun;Ko, Ha-Yoon;Baek, Joong-Hwan
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.22 no.3
    • /
    • pp.471-479
    • /
    • 2018
  • Recently, the spread of smart TV based Android iOS Set Top box has become common. This paper propose a new approach to control the TV using gestures away from the era of controlling the TV using remote control. In this paper, the AdaBoost algorithm is applied to gesture recognition by using a mono camera. First, we use Camshift-based Body tracking and estimation algorithm based on Gaussian background removal for body coordinate extraction. Using global and local feature vectors, we recognized gestures with speed change. By tracking the time interval trajectories of hand and wrist, the AdaBoost algorithm with CART algorithm is used to train and classify gestures. The principal component feature vector with high classification success rate is searched using CART algorithm. As a result, 24 optimal feature vectors were found, which showed lower error rate (3.73%) and higher accuracy rate (95.17%) than the existing algorithm.

Hand Motion Signal Extraction Based on Electric Field Sensors Using PLN Spectrum Analysis (PLN 성분 분석을 통한 전기장센서 기반 손동작신호 추출)

  • Jeong, Seonil;Kim, Youngchul
    • Smart Media Journal
    • /
    • v.9 no.4
    • /
    • pp.97-101
    • /
    • 2020
  • Using passive electric field sensor which operates in non-contact mode, we can measure the electric potential induced from the change of electric charges on a sensor caused by the movement of human body or hands. In this study, we propose a new method, which utilizes PLN induced to the sensor around the moving object, to detect one's hand movement and extract gesture frames from the detected signals. Signals from the EPS sensors include a large amount of power line noise usually existing in the places such as rooms or buildings. Using the fact that the PLN is shielded in part by human access to the sensor, signals caused by motion or hand movement are detected. PLN consists mainly of signals with frequency of 60 Hz and its harmonics. In our proposed method, signals only 120 Hz component in frequency domain are chosen selectively and exclusively utilized for detection of hand movement. We use FFT to measure a spectral-separated frequency signal. The signals obtained from sensors in this way are continued to be compared with the threshold preset in advance. Once motion signals are detected passing throng the threshold, we determine the motion frame based on period between the first threshold passing time and the last one. The motion detection rate of our proposed method was about 90% while the correct frame extraction rate was about 85%. The method like our method, which use PLN signal in order to extract useful data about motion movement from non-contact mode EPS sensors, has been rarely reported or published in recent. This research results can be expected to be useful especially in circumstance of having surrounding PLN.

Automatic Coarticulation Detection for Continuous Sign Language Recognition (연속된 수화 인식을 위한 자동화된 Coarticulation 검출)

  • Yang, Hee-Deok;Lee, Seong-Whan
    • Journal of KIISE:Software and Applications
    • /
    • v.36 no.1
    • /
    • pp.82-91
    • /
    • 2009
  • Sign language spotting is the task of detecting and recognizing the signs in a signed utterance. The difficulty of sign language spotting is that the occurrences of signs vary in both motion and shape. Moreover, the signs appear within a continuous gesture stream, interspersed with transitional movements between signs in a vocabulary and non-sign patterns(which include out-of-vocabulary signs, epentheses, and other movements that do not correspond to signs). In this paper, a novel method for designing a threshold model in a conditional random field(CRF) model is proposed. The proposed model performs an adaptive threshold for distinguishing between signs in the vocabulary and non-sign patterns. A hand appearance-based sign verification method, a short-sign detector, and a subsign reasoning method are included to further improve sign language spotting accuracy. Experimental results show that the proposed method can detect signs from continuous data with an 88% spotting rate and can recognize signs from isolated data with a 94% recognition rate, versus 74% and 90% respectively for CRFs without a threshold model, short-sign detector, subsign reasoning, and hand appearance-based sign verification.

Study on Signal Processing Method for Extracting Hand-Gesture Signals Using Sensors Measuring Surrounding Electric Field Disturbance (주변 전기장 측정센서를 이용한 손동작 신호 검출을 위한 신호처리시스템 연구)

  • Cheon, Woo Young;Kim, Young Chul
    • Smart Media Journal
    • /
    • v.6 no.2
    • /
    • pp.26-32
    • /
    • 2017
  • In this paper, we implement a signal-detecting electric circuit based LED lighting control system which is essential in NUI technology using EPIC converting surrounding earth electric field disturbance signals to electric potential signals. We used signal-detecting electric circuits which was developed to extract individual signal for each EPIC sensor while conventional EPIC-based development equipments provide limited forms of signals. The signals extracted from our developed circuit contributed to better performance as well as flexiblity in processes of feature extracting stage and pattern recognition stage. We designed a system which can control the brightness and on/off of LED lights with four hand gestures in order to justify its applicability to real application systems. We obtained faster pattern classification speed not only by developing an instruction system, but also by using interface control signals.