• Title/Summary/Keyword: Gesture segmentation

Search Result 41, Processing Time 0.024 seconds

Hand Gesture Segmentation Method using a Wrist-Worn Wearable Device

  • Lee, Dong-Woo;Son, Yong-Ki;Kim, Bae-Sun;Kim, Minkyu;Jeong, Hyun-Tae;Cho, Il-Yeon
    • Journal of the Ergonomics Society of Korea
    • /
    • v.34 no.5
    • /
    • pp.541-548
    • /
    • 2015
  • Objective: We introduce a hand gesture segmentation method using a wrist-worn wearable device which can recognize simple gestures of clenching and unclenching ones' fist. Background: There are many types of smart watches and fitness bands in the markets. And most of them already adopt a gesture interaction to provide ease of use. However, there are many cases in which the malfunction is difficult to distinguish between the user's gesture commands and user's daily life motion. It is needed to develop a simple and clear gesture segmentation method to improve the gesture interaction performance. Method: At first, we defined the gestures of making a fist (start of gesture command) and opening one's fist (end of gesture command) as segmentation gestures to distinguish a gesture. The gestures of clenching and unclenching one's fist are simple and intuitive. And we also designed a single gesture consisting of a set of making a fist, a command gesture, and opening one's fist in order. To detect segmentation gestures at the bottom of the wrist, we used a wrist strap on which an array of infrared sensors (emitters and receivers) were mounted. When a user takes gestures of making a fist and opening one's a fist, this changes the shape of the bottom of the wrist, and simultaneously changes the reflected amount of the infrared light detected by the receiver sensor. Results: An experiment was conducted in order to evaluate gesture segmentation performance. 12 participants took part in the experiment: 10 males, and 2 females with an average age of 38. The recognition rates of the segmentation gestures, clenching and unclenching one's fist, are 99.58% and 100%, respectively. Conclusion: Through the experiment, we have evaluated gesture segmentation performance and its usability. The experimental results show a potential for our suggested segmentation method in the future. Application: The results of this study can be used to develop guidelines to prevent injury in auto workers at mission assembly plants.

A Memory-efficient Hand Segmentation Architecture for Hand Gesture Recognition in Low-power Mobile Devices

  • Choi, Sungpill;Park, Seongwook;Yoo, Hoi-Jun
    • JSTS:Journal of Semiconductor Technology and Science
    • /
    • v.17 no.3
    • /
    • pp.473-482
    • /
    • 2017
  • Hand gesture recognition is regarded as new Human Computer Interaction (HCI) technologies for the next generation of mobile devices. Previous hand gesture implementation requires a large memory and computation power for hand segmentation, which fails to give real-time interaction with mobile devices to users. Therefore, in this paper, we presents a low latency and memory-efficient hand segmentation architecture for natural hand gesture recognition. To obtain both high memory-efficiency and low latency, we propose a streaming hand contour tracing unit and a fast contour filling unit. As a result, it achieves 7.14 ms latency with only 34.8 KB on-chip memory, which are 1.65 times less latency and 1.68 times less on-chip memory, respectively, compare to the best-in-class.

A study on hand gesture recognition using 3D hand feature (3차원 손 특징을 이용한 손 동작 인식에 관한 연구)

  • Bae Cheol-Soo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.10 no.4
    • /
    • pp.674-679
    • /
    • 2006
  • In this paper a gesture recognition system using 3D feature data is described. The system relies on a novel 3D sensor that generates a dense range mage of the scene. The main novelty of the proposed system, with respect to other 3D gesture recognition techniques, is the capability for robust recognition of complex hand postures such as those encountered in sign language alphabets. This is achieved by explicitly employing 3D hand features. Moreover, the proposed approach does not rely on colour information, and guarantees robust segmentation of the hand under various illumination conditions, and content of the scene. Several novel 3D image analysis algorithms are presented covering the complete processing chain: 3D image acquisition, arm segmentation, hand -forearm segmentation, hand pose estimation, 3D feature extraction, and gesture classification. The proposed system is tested in an application scenario involving the recognition of sign-language postures.

Proposal of Image Detection Algorithm to Implement Hand Gestures

  • Woo, Eun-Ju;Moon, Yu-Sung;Choi, Ung-Se;Kim, Jung-Won
    • Journal of IKEEE
    • /
    • v.22 no.4
    • /
    • pp.1222-1225
    • /
    • 2018
  • This paper proposes an image detection algorithm to implement gesture. By using a camera sensor, the performance of the extracted image algorithm based on the gesture pattern was verified through experiments. In addition, through the experiments, we confirmed the proposed method's possibility of the implementation. For efficient image detection, we applied a segmentation technique based on image transition which divides into small units. To improve gesture recognition, the proposed method not only has high recognition rate and low false acceptance rate in real gesture environment, but also designed an algorithm that efficiently finds optimal thresholds that can be applied.

Interactive visual knowledge acquisition for hand-gesture recognition (손 제스쳐 인식을 위한 상호작용 시각정보 추출)

  • 양선옥;최형일
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.33B no.9
    • /
    • pp.88-96
    • /
    • 1996
  • Computer vision-based gesture recognition systems consist of image segmentation, object tracking and decision. However, it is difficult to segment an object from image for gesture in computer systems because of vaious illuminations and backgrounds. In this paper, we describe a method to learn features for segmentation, which improves the performance of computer vision-based hand-gesture recognition systems. Systems interact with a user to acquire exact training data and segment information according to a predefined plan. System provides some models to the user, takes pictures of the user's response and then analyzes the pictures with models and a prior knowledge. The system sends messages to the user and operates learning module to extract information with the analyzed result.

  • PDF

Hybrid HMM for Transitional Gesture Classification in Thai Sign Language Translation

  • Jaruwanawat, Arunee;Chotikakamthorn, Nopporn;Werapan, Worawit
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2004.08a
    • /
    • pp.1106-1110
    • /
    • 2004
  • A human sign language is generally composed of both static and dynamic gestures. Each gesture is represented by a hand shape, its position, and hand movement (for a dynamic gesture). One of the problems found in automated sign language translation is on segmenting a hand movement that is part of a transitional movement from one hand gesture to another. This transitional gesture conveys no meaning, but serves as a connecting period between two consecutive gestures. Based on the observation that many dynamic gestures as appeared in Thai sign language dictionary are of quasi-periodic nature, a method was developed to differentiate between a (meaningful) dynamic gesture and a transitional movement. However, there are some meaningful dynamic gestures that are of non-periodic nature. Those gestures cannot be distinguished from a transitional movement by using the signal quasi-periodicity. This paper proposes a hybrid method using a combination of the periodicity-based gesture segmentation method with a HMM-based gesture classifier. The HMM classifier is used here to detect dynamic signs of non-periodic nature. Combined with the periodic-based gesture segmentation method, this hybrid scheme can be used to identify segments of a transitional movement. In addition, due to the use of quasi-periodic nature of many dynamic sign gestures, dimensionality of the HMM part of the proposed method is significantly reduced, resulting in computational saving as compared with a standard HMM-based method. Through experiment with real measurement, the proposed method's recognition performance is reported.

  • PDF

Controlling Slides using Hand tracking and Gesture Recognition (손의 추적과 제스쳐 인식에 의한 슬라이드 제어)

  • Fayyaz, Rabia;Rhee, Eun Joo
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2012.04a
    • /
    • pp.436-439
    • /
    • 2012
  • The work is to the control the desktop Computers based on hand gesture recognition. This paper is worked en real time tracking and recognizes the hand gesture for controlling the slides based on hand direction such as right and left using a real time camera.

Dynamic Gesture Recognition for the Remote Camera Robot Control (원격 카메라 로봇 제어를 위한 동적 제스처 인식)

  • Lee Ju-Won;Lee Byung-Ro
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.8 no.7
    • /
    • pp.1480-1487
    • /
    • 2004
  • This study is proposed the novel gesture recognition method for the remote camera robot control. To recognize the dynamics gesture, the preprocessing step is the image segmentation. The conventional methods for the effectively object segmentation has need a lot of the cole. information about the object(hand) image. And these methods in the recognition step have need a lot of the features with the each object. To improve the problems of the conventional methods, this study proposed the novel method to recognize the dynamic hand gesture such as the MMS(Max-Min Search) method to segment the object image, MSM(Mean Space Mapping) method and COG(Conte. Of Gravity) method to extract the features of image, and the structure of recognition MLPNN(Multi Layer Perceptron Neural Network) to recognize the dynamic gestures. In the results of experiment, the recognition rate of the proposed method appeared more than 90[%], and this result is shown that is available by HCI(Human Computer Interface) device for .emote robot control.

Design and Implementation for Korean Character and Pen-gesture Recognition System using Stroke Information (획 정보를 이용한 한글문자와 펜 제스처 인식 시스템의 설계 및 구현)

  • Oh, Jun-Taek;Kim, Wook-Hyun
    • The KIPS Transactions:PartB
    • /
    • v.9B no.6
    • /
    • pp.765-774
    • /
    • 2002
  • The purpose of this paper is a design and implementation for korean character and pen-gesture recognition system in multimedia terminal, PDA and etc, which demand both a fast process and a high recognition rate. To recognize writing-types which are written by various users, the korean character recognition system uses a database which is based on the characteristic information of korean and the stroke information Which composes a phoneme, etc. In addition. it has a fast speed by the phoneme segmentation which uses the successive process or the backtracking process. The pen-gesture recognition system is performed by a matching process between the classification features extracted from an input pen-gesture and the classification features of 15 pen-gestures types defined in the gesture model. The classification feature is using the insensitive stroke information. i.e., the positional relation between two strokes. the crossing number, the direction transition, the direction vector, the number of direction code. and the distance ratio between starting and ending point in each stroke. In the experiment, we acquired a high recognition rate and a fart speed.

Palm Area Detection by Maximum Hand Width (손 최장너비 기반 손바닥 영역 검출)

  • Choi, Eun Chang;Kim, Jun Yeon;Lee, Jae Won;Lim, Jong Gwan
    • The Journal of the Korea Contents Association
    • /
    • v.18 no.4
    • /
    • pp.398-405
    • /
    • 2018
  • In the HCI, hand gesture recognition is attracting attention as a method for interaction and information exchange between users and devices along with the development of IT devices. In hand gesture recognition through image processing, palm region detection is a key process contributing to improvement of processing speed and recognition rate. In this paper, we propose a new method for image segmentation between the hand and wrist for palm area detection. The anatomical characteristics of the hand are used to calculate the distance between the iliac bones of the thumb and little finger, which have the widest width, by the horizontal projection histogram of the hand image, and then the palm area is detected by drawing a circle having the width as the diameter. In order to verify the superiority of this method, multiple stage template matching is used to compare and evaluate recognition performance against the four conventional methods for 10 hand gestures. Note that the literatures to offer palm area detection performance evaluation are few although there are many studies on hand gesture recognition.