• Title/Summary/Keyword: Direction recognition

Search Result 848, Processing Time 0.026 seconds

Recognition Direction Improvement of Target Object for Machine Vision based Automatic Inspection (머신비전 자동검사를 위한 대상객체의 인식방향성 개선)

  • Hong, Seung-Beom;Hong, Seung-Woo;Lee, Kyou-Ho
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.23 no.11
    • /
    • pp.1384-1390
    • /
    • 2019
  • This paper proposes a technological solution for improving the recognition direction of target objects for automatic vision inspection by machine vision. This paper proposes a technological solution for improving the recognition direction of target objects for automatic vision inspection by machine vision. This enables the automatic machine vision inspection to detect the image of the inspection object regardless of the position and orientation of the object, eliminating the need for a separate inspection jig and improving the automation level of the inspection process. This study develops the technology and method that can be applied to the wire harness manufacturing process as the inspection object and present the result of real system. The results of the system implementation was evaluated by the accredited institution. This includes successful measurement in the accuracy, detection recognition, reproducibility and positioning success rate, and achievement the goal in ten kinds of color discrimination ability, inspection time within one second and four automatic mode setting, etc.

Design and Implementation for Korean Character and Pen-gesture Recognition System using Stroke Information (획 정보를 이용한 한글문자와 펜 제스처 인식 시스템의 설계 및 구현)

  • Oh, Jun-Taek;Kim, Wook-Hyun
    • The KIPS Transactions:PartB
    • /
    • v.9B no.6
    • /
    • pp.765-774
    • /
    • 2002
  • The purpose of this paper is a design and implementation for korean character and pen-gesture recognition system in multimedia terminal, PDA and etc, which demand both a fast process and a high recognition rate. To recognize writing-types which are written by various users, the korean character recognition system uses a database which is based on the characteristic information of korean and the stroke information Which composes a phoneme, etc. In addition. it has a fast speed by the phoneme segmentation which uses the successive process or the backtracking process. The pen-gesture recognition system is performed by a matching process between the classification features extracted from an input pen-gesture and the classification features of 15 pen-gestures types defined in the gesture model. The classification feature is using the insensitive stroke information. i.e., the positional relation between two strokes. the crossing number, the direction transition, the direction vector, the number of direction code. and the distance ratio between starting and ending point in each stroke. In the experiment, we acquired a high recognition rate and a fart speed.

Automatic Recognition of Direction Information in Road Sign Image Using OpenCV (OpenCV를 이용한 도로표지 영상에서의 방향정보 자동인식)

  • Kim, Gihong;Chong, Kyusoo;Youn, Junhee
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.31 no.4
    • /
    • pp.293-300
    • /
    • 2013
  • Road signs are important infrastructures for safe and smooth traffic by providing useful information to drivers. It is necessary to establish road sign DB for managing road signs systematically. To provide such DB, manually detection and recognition from imagery can be done. However, it is time and cost consuming. In this study, we proposed algorithms for automatic recognition of direction information in road sign image. Also we developed algorithm code using OpenCV library, and applied it to road sign image. To automatically detect and recognize direction information, we developed program which is composed of various modules such as image enhancement, image binarization, arrow region extraction, interesting point extraction, and template image matching. As a result, we can confirm the possibility of automatic recognition of direction information in road sign image.

Study on Design of Fingerprint Recognition Embedded System using Neural Network (신경망을 이용한 지문인식 임베디드 시스템 설계에 관한 연구)

  • Lee Jae-Hyun;Kim Dong-Han
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.10 no.4
    • /
    • pp.775-782
    • /
    • 2006
  • We generated blocks from the direction-extracted fingerprint during the pre-process of the fingerprint recognition algorithm and performed training by using the direction minutiae of each block as the input pattern of the neural network, so that we extracted the core points to use in the matching. Based on this, we designed the fingerprint recognition embedded system and tested it using the control board and the serial communication to utilize it for a variety of application systems. As a result, we can verify the reliance satisfactorily.

Study on Design of Fingerprint Recognition Embedded System using Neural Network

  • Kim, Dong Han;Kim, Jung Hoon;Lee, Sang Hae
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.4 no.3
    • /
    • pp.347-352
    • /
    • 2004
  • We generated blocks from the direction-extracted fingerprint during the pre-process of the fingerprint recognition algorithm and performed training by using the direction minutiae of each block as the input pattern of the neural network, so that we extracted the core points to use in the matching. Based on this, we designed the fingerprint recognition embedded system and tested it by using the control board and the serial communication to utilize it for a variety of application systems. As a result, we can verify the reliance satisfactorily.

An Implementation of Gaze Direction Recognition System using Difference Image Entropy (차영상 엔트로피를 이용한 시선 인식 시스템의 구현)

  • Lee, Kue-Bum;Chung, Dong-Keun;Hong, Kwang-Seok
    • The KIPS Transactions:PartB
    • /
    • v.16B no.2
    • /
    • pp.93-100
    • /
    • 2009
  • In this paper, we propose a Difference Image Entropy based gaze direction recognition system. The Difference Image Entropy is computed by histogram levels using the acquired difference image of current image and reference images or average images that have peak positions from $-255{\sim}+255$ to prevent information omission. There are two methods about the Difference Image Entropy based gaze direction. 1) The first method is to compute the Difference Image Entropy between an input image and average images of 45 images in each location of gaze, and to recognize the directions of user's gaze. 2) The second method is to compute the Difference Image Entropy between an input image and each 45 reference images, and to recognize the directions of user's gaze. The reference image is created by average image of 45 images in each location of gaze after receiving images of 4 directions. In order to evaluate the performance of the proposed system, we conduct comparison experiment with PCA based gaze direction system. The directions of recognition left-top, right-top, left-bottom, right-bottom, and we make an experiment on that, as changing the part of recognition about 45 reference images or average image. The experimental result shows that the recognition rate of Difference Image Entropy is 97.00% and PCA is 95.50%, so the recognition rate of Difference Image Entropy based system is 1.50% higher than PCA based system.

Human Activity Recognition Using Spatiotemporal 3-D Body Joint Features with Hidden Markov Models

  • Uddin, Md. Zia;Kim, Jaehyoun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.6
    • /
    • pp.2767-2780
    • /
    • 2016
  • Video-based human-activity recognition has become increasingly popular due to the prominent corresponding applications in a variety of fields such as computer vision, image processing, smart-home healthcare, and human-computer interactions. The essential goals of a video-based activity-recognition system include the provision of behavior-based information to enable functionality that proactively assists a person with his/her tasks. The target of this work is the development of a novel approach for human-activity recognition, whereby human-body-joint features that are extracted from depth videos are used. From silhouette images taken at every depth, the direction and magnitude features are first obtained from each connected body-joint pair so that they can be augmented later with motion direction, as well as with the magnitude features of each joint in the next frame. A generalized discriminant analysis (GDA) is applied to make the spatiotemporal features more robust, followed by the feeding of the time-sequence features into a Hidden Markov Model (HMM) for the training of each activity. Lastly, all of the trained-activity HMMs are used for depth-video activity recognition.

Gaze Recognition Interface Development for Smart Wheelchair (지능형 휠체어를 위한 시선 인식 인터페이스 개발)

  • Park, S.H.
    • Journal of rehabilitation welfare engineering & assistive technology
    • /
    • v.5 no.1
    • /
    • pp.103-110
    • /
    • 2011
  • In this paper, we propose a gaze recognition interface for smart wheelchair. The gaze recognition interface is a user interface which recognize the commands using the gaze recognition and avoid the detected obstacles by sensing the distance through range sensors on the way to driving. Smart wheelchair is composed of gaze recognition and tracking module, user interface module, obstacle detector, motor control module, and range sensor module. The interface in this paper uses a camera with built-in infra red filter and 2 LED light sources to see what direction the pupils turn to and can send command codes to control the system, thus it doesn't need any correction process per each person. The results of the experiment showed that the proposed interface can control the system exactly by recognizing user's gaze direction.

A Study on On-line Recognition System of Korean Characters (온라인 한글자소 인식시스템의 구성에 관한 연구)

  • Choi, Seok;Kim, Gil-Jung;Huh, Man-Tak;Lee, Jong-Hyeok;Nam, Ki-Gon;Yoon, Tae-Hoon;Kim, Jae-Chang;Lee, Ryang-Seong
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.30B no.9
    • /
    • pp.94-105
    • /
    • 1993
  • In this paper propose a Koaren character recognition system using a neural network is proposed. This system is a multilayer neural network based on the masking field model which consists of a input layer, four feature extraction layers which extracts type, direction, stroke, and connection features, and an output layer which gives us recognized character codes. First, 4x4 subpatterns of an NxN character pattern stored in the input buffer are applied into the feature extraction layers sequentially. Then, each of feature extraction layers extracts sequentially features such as type, direction, stroke, and connection, respectively. Type features for direction and connection are extracted by the type feature extraction layer, direction features for stroke by the direction feature extraction layer and stroke and connection features for stroke by the direction feature extraction layer and stroke and connection features for the recongnition of character by the stroke and the connection feature extractions layers, respectively. The stroke and connection features are saved in the sequential buffer layer sequentially and using these features the characters are recognized in the output layer. The recognition results of this system by tests with 8 single consonants and 6 single vowels are promising.

  • PDF

Recognition of Hmm Facial Expressions using Optical Flow of Feature Regions (얼굴 특징영역상의 광류를 이용한 표정 인식)

  • Lee Mi-Ae;Park Ki-Soo
    • Journal of KIISE:Software and Applications
    • /
    • v.32 no.6
    • /
    • pp.570-579
    • /
    • 2005
  • Facial expression recognition technology that has potentialities for applying various fields is appling on the man-machine interface development, human identification test, and restoration of facial expression by virtual model etc. Using sequential facial images, this study proposes a simpler method for detecting human facial expressions such as happiness, anger, surprise, and sadness. Moreover the proposed method can detect the facial expressions in the conditions of the sequential facial images which is not rigid motion. We identify the determinant face and elements of facial expressions and then estimates the feature regions of the elements by using information about color, size, and position. In the next step, the direction patterns of feature regions of each element are determined by using optical flows estimated gradient methods. Using the direction model proposed by this study, we match each direction patterns. The method identifies a facial expression based on the least minimum score of combination values between direction model and pattern matching for presenting each facial expression. In the experiments, this study verifies the validity of the Proposed methods.