• Title/Summary/Keyword: Vision-based interface

Search Result 131, Processing Time 0.023 seconds

Real Time Gaze Discrimination for Computer Interface (컴퓨터 인터페이스를 위한 실시간 시선 식별)

  • Hwang, Suen-Ki;Kim, Moon-Hwan
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.3 no.1
    • /
    • pp.38-46
    • /
    • 2010
  • This paper describes a computer vision system based on active IR illumination for real-time gaze discrimination system. Unlike most of the existing gaze discrimination techniques, which often require assuming a static head to work well and require a cumbersome calibration process for each person, our gaze discrimination system can perform robust and accurate gaze estimation without calibration and under rather significant head movement. This is made possible by a new gaze calibration procedure that identifies the mapping from pupil parameters to screen coordinates using generalized regression neural networks (GRNNs). With GRNNs, the mapping does not have to be an analytical function and head movement is explicitly accounted for by the gaze mapping function. Furthermore, the mapping function can generalize to other individuals not used in the training. To further improve the gaze estimation accuracy, we employ a reclassification scheme that deals with the classes that tend to be misclassified. This leads to a 10% improvement in classification error. The angular gaze accuracy is about $5^{\circ}$horizontally and $8^{\circ}$vertically. The effectiveness of our gaze tracker is demonstrated by experiments that involve gaze-contingent interactive graphic display.

  • PDF

Automatic Detection of Dissimilar Regions through Multiple Feature Analysis (다중의 특징 분석을 통한 비 유사 영역의 자동적인 검출)

  • Jang, Seok-Woo;Jung, Myunghee
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.2
    • /
    • pp.160-166
    • /
    • 2020
  • As mobile-based hardware technology develops, many kinds of applications are also being developed. In addition, there is an increasing demand to automatically check that the interface of these applications works correctly. In this paper, we describe a method for accurately detecting faulty images from applications by comparing major characteristics from input color images. For this purpose, our method first extracts major characteristics of the input image, then calculates the differences in the extracted major features, and decides if the test image is a normal image or a faulty image dissimilar to the reference image. Experiment results show that the suggested approach robustly determines similar and dissimilar images by comparing major characteristics from input color images. The suggested method is expected to be useful in many real application areas related to computer vision, like video indexing, object detection and tracking, image surveillance, and so on.

A Study on Hand Gesture Recognition with Low-Resolution Hand Images (저해상도 손 제스처 영상 인식에 대한 연구)

  • Ahn, Jung-Ho
    • Journal of Satellite, Information and Communications
    • /
    • v.9 no.1
    • /
    • pp.57-64
    • /
    • 2014
  • Recently, many human-friendly communication methods have been studied for human-machine interface(HMI) without using any physical devices. One of them is the vision-based gesture recognition that this paper deals with. In this paper, we define some gestures for interaction with objects in a predefined virtual world, and propose an efficient method to recognize them. For preprocessing, we detect and track the both hands, and extract their silhouettes from the low-resolution hand images captured by a webcam. We modeled skin color by two Gaussian distributions in RGB color space and use blob-matching method to detect and track the hands. Applying the foodfill algorithm we extracted hand silhouettes and recognize the hand shapes of Thumb-Up, Palm and Cross by detecting and analyzing their modes. Then, with analyzing the context of hand movement, we recognized five predefined one-hand or both-hand gestures. Assuming that one main user shows up for accurate hand detection, the proposed gesture recognition method has been proved its efficiency and accuracy in many real-time demos.

Real-Time Object Recognition Using Local Features (지역 특징을 사용한 실시간 객체인식)

  • Kim, Dae-Hoon;Hwang, Een-Jun
    • Journal of IKEEE
    • /
    • v.14 no.3
    • /
    • pp.224-231
    • /
    • 2010
  • Automatic detection of objects in images has been one of core challenges in the areas such as computer vision and pattern analysis. Especially, with the recent deployment of personal mobile devices such as smart phone, such technology is required to be transported to them. Usually, these smart phone users are equipped with devices such as camera, GPS, and gyroscope and provide various services through user-friendly interface. However, the smart phones fail to give excellent performance due to limited system resources. In this paper, we propose a new scheme to improve object recognition performance based on pre-computation and simple local features. In the pre-processing, we first find several representative parts from similar type objects and classify them. In addition, we extract features from each classified part and train them using regression functions. For a given query image, we first find candidate representative parts and compare them with trained information to recognize objects. Through experiments, we have shown that our proposed scheme can achieve resonable performance.

Conceptual Study on Coaxial Rotorcraft UAV for teaming operation with UGV (무인지상차량과의 합동운용을 위한 동축반전 회전익형 무인항공기 개념연구)

  • Byun, Young-Seop;Song, Jun-Beom;Song, Woo-Jin;Kim, Jeong;Kang, Beom-Soo
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.39 no.5
    • /
    • pp.458-465
    • /
    • 2011
  • UAV-UGV teaming concept has been proposed that can compensate for weak points of each platform by providing carrying, launching, recovery and recharging capability for the VTOL-UAV through the host UGV. The teaming concept can expand the observation envelop of the UGV and extend the operational capability of the UAV through mechanical combination of each system. The spherical-shaped coaxial rotorcraft UAV is suggested to provide flexible and precise interface between two systems. Hybrid navigation solution that included vision-based target tracking method for precision landing is investigated and its experimental study is performed. Feasibility study on length-variable rotor to provide the compact configuration of the loaded rotorcraft platform is also described.

Hand Gesture Recognition using Multivariate Fuzzy Decision Tree and User Adaptation (다변량 퍼지 의사결정트리와 사용자 적응을 이용한 손동작 인식)

  • Jeon, Moon-Jin;Do, Jun-Hyeong;Lee, Sang-Wan;Park, Kwang-Hyun;Bien, Zeung-Nam
    • The Journal of Korea Robotics Society
    • /
    • v.3 no.2
    • /
    • pp.81-90
    • /
    • 2008
  • While increasing demand of the service for the disabled and the elderly people, assistive technologies have been developed rapidly. The natural signal of human such as voice or gesture has been applied to the system for assisting the disabled and the elderly people. As an example of such kind of human robot interface, the Soft Remote Control System has been developed by HWRS-ERC in $KAIST^[1]$. This system is a vision-based hand gesture recognition system for controlling home appliances such as television, lamp and curtain. One of the most important technologies of the system is the hand gesture recognition algorithm. The frequently occurred problems which lower the recognition rate of hand gesture are inter-person variation and intra-person variation. Intra-person variation can be handled by inducing fuzzy concept. In this paper, we propose multivariate fuzzy decision tree(MFDT) learning and classification algorithm for hand motion recognition. To recognize hand gesture of a new user, the most proper recognition model among several well trained models is selected using model selection algorithm and incrementally adapted to the user's hand gesture. For the general performance of MFDT as a classifier, we show classification rate using the benchmark data of the UCI repository. For the performance of hand gesture recognition, we tested using hand gesture data which is collected from 10 people for 15 days. The experimental results show that the classification and user adaptation performance of proposed algorithm is better than general fuzzy decision tree.

  • PDF

A Method of Hand Recognition for Virtual Hand Control of Virtual Reality Game Environment (가상 현실 게임 환경에서의 가상 손 제어를 위한 사용자 손 인식 방법)

  • Kim, Boo-Nyon;Kim, Jong-Ho;Kim, Tae-Young
    • Journal of Korea Game Society
    • /
    • v.10 no.2
    • /
    • pp.49-56
    • /
    • 2010
  • In this paper, we propose a control method of virtual hand by the recognition of a user's hand in the virtual reality game environment. We display virtual hand on the game screen after getting the information of the user's hand movement and the direction thru input images by camera. We can utilize the movement of a user's hand as an input interface for virtual hand to select and move the object. As a hand recognition method based on the vision technology, the proposed method transforms input image from RGB color space to HSV color space, then segments the hand area using double threshold of H, S value and connected component analysis. Next, The center of gravity of the hand area can be calculated by 0 and 1 moment implementation of the segmented area. Since the center of gravity is positioned onto the center of the hand, the further apart pixels from the center of the gravity among the pixels in the segmented image can be recognized as fingertips. Finally, the axis of the hand is obtained as the vector of the center of gravity and the fingertips. In order to increase recognition stability and performance the method using a history buffer and a bounding box is also shown. The experiments on various input images show that our hand recognition method provides high level of accuracy and relatively fast stable results.

The Present Status and Vision of Virtual Construction System Development (가상 건설 시스템 개발 현황과 비전)

  • Kim, Jae-Jun;Choi, Cul-Ho;Shin, Hyun-Mok;Jin, Sang-Yoon;Lee, Kwang-Myung
    • Proceedings of the Korean Institute Of Construction Engineering and Management
    • /
    • 2008.11a
    • /
    • pp.170-175
    • /
    • 2008
  • The research team for the virtual construction development was established with the support of Korea Ministry of Construction and Transportation, and KICTEP (Korea Institute of Construction and Transportation Technology Evaluation and Planning). Its aims are to develop system that is to improve productivity & quality, to create a higher value-added business, and to cultivate international competitiveness in the construction industry. The virtual construction system is a design, engineering, and construction management information system that allows the project participants to effectively share the information throughout the construction life cycle with the support of 3D and design information. To achieve this, the research team focuses on developing several systems. First, the team focuses on developing for the pre-planning, the structural engineering, MEP, and the 3D based estimation system. Second, they focus on developing a simulation system for the construction process planning and feasibility study with help of the virtual reality technologies. Third, they focus on developing the CPLM (Construction Project Life-cycle Management) system for managing construction project data, and the decision support system that makes the collaboration among the project participants based on 3D technologies and information. We also focus on developing the SDAI (Standard Data Access Interface), the localized guideline for 3D design, and a training program. In addition, we focus on developing the undeveloped area of the commercial system and building an environment that can support the communication and collaboration in the construction life-cycle rather than developing the existing and commercialized system.

  • PDF

Multi-classifier Decision-level Fusion for Face Recognition (다중 분류기의 판정단계 융합에 의한 얼굴인식)

  • Yeom, Seok-Won
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.49 no.4
    • /
    • pp.77-84
    • /
    • 2012
  • Face classification has wide applications in intelligent video surveillance, content retrieval, robot vision, and human-machine interface. Pose and expression changes, and arbitrary illumination are typical problems for face recognition. When the face is captured at a distance, the image quality is often degraded by blurring and noise corruption. This paper investigates the efficacy of multi-classifier decision level fusion for face classification based on the photon-counting linear discriminant analysis with two different cost functions: Euclidean distance and negative normalized correlation. Decision level fusion comprises three stages: cost normalization, cost validation, and fusion rules. First, the costs are normalized into the uniform range and then, candidate costs are selected during validation. Three fusion rules are employed: minimum, average, and majority-voting rules. In the experiments, unfocusing and motion blurs are rendered to simulate the effects of the long distance environments. It will be shown that the decision-level fusion scheme provides better results than the single classifier.

Design and Implementation of a Hardware Accelerator for Marine Object Detection based on a Binary Segmentation Algorithm for Ship Safety Navigation (선박안전 운항을 위한 이진 분할 알고리즘 기반 해상 객체 검출 하드웨어 가속기 설계 및 구현)

  • Lee, Hyo-Chan;Song, Hyun-hak;Lee, Sung-ju;Jeon, Ho-seok;Kim, Hyo-Sung;Im, Tae-ho
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.10
    • /
    • pp.1331-1340
    • /
    • 2020
  • Object detection in maritime means that the captain detects floating objects that has a risk of colliding with the ship using the computer automatically and as accurately as human eyes. In conventional ships, the presence and distance of objects are determined through radar waves. However, it cannot identify the shape and type. In contrast, with the development of AI, cameras help accurately identify obstacles on the sea route with excellent performance in detecting or recognizing objects. The computer must calculate high-volume pixels to analyze digital images. However, the CPU is specialized for sequential processing; the processing speed is very slow, and smooth service support or security is not guaranteed. Accordingly, this study developed maritime object detection software and implemented it with FPGA to accelerate the processing of large-scale computations. Additionally, the system implementation was improved through embedded boards and FPGA interface, achieving 30 times faster performance than the existing algorithm and a three-times faster entire system.