• Title/Summary/Keyword: Vision Detection

Search Result 1,292, Processing Time 0.023 seconds

A Vision-Based Method to Find Fingertips in a Closed Hand

  • Chaudhary, Ankit;Vatwani, Kapil;Agrawal, Tushar;Raheja, J.L.
    • Journal of Information Processing Systems
    • /
    • v.8 no.3
    • /
    • pp.399-408
    • /
    • 2012
  • Hand gesture recognition is an important area of research in the field of Human Computer Interaction (HCI). The geometric attributes of the hand play an important role in hand shape reconstruction and gesture recognition. That said, fingertips are one of the important attributes for the detection of hand gestures and can provide valuable information from hand images. Many methods are available in scientific literature for fingertips detection with an open hand but very poor results are available for fingertips detection when the hand is closed. This paper presents a new method for the detection of fingertips in a closed hand using the corner detection method and an advanced edge detection algorithm. It is important to note that the skin color segmentation methodology did not work for fingertips detection in a closed hand. Thus the proposed method applied Gabor filter techniques for the detection of edges and then applied the corner detection algorithm for the detection of fingertips through the edges. To check the accuracy of the method, this method was tested on a vast number of images taken with a webcam. The method resulted in a higher accuracy rate of detections from the images. The method was further implemented on video for testing its validity on real time image capturing. These closed hand fingertips detection would help in controlling an electro-mechanical robotic hand via hand gesture in a natural way.

Evolutionary Generation Based Color Detection Technique for Object Identification in Degraded Robot Vision (저하된 로봇 비전에서의 물체 인식을 위한 진화적 생성 기반의 컬러 검출 기법)

  • Kim, Kyoungtae;Seo, Kisung
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.64 no.7
    • /
    • pp.1040-1046
    • /
    • 2015
  • This paper introduces GP(Genetic Programming) based color detection model for an object detection of humanoid robot vision. Existing color detection methods have used linear/nonlinear transformation of RGB color-model. However, most of cases have difficulties to classify colors satisfactory because of interference of among color channels and susceptibility for illumination variation. Especially, they are outstanding in degraded images from robot vision. To solve these problems, we propose illumination robust and non-parametric multi-colors detection model using evolution of GP. The proposed method is compared to the existing color-models for various environments in robot vision for real humanoid Nao.

Investigation on the Real-Time Environment Recognition System Based on Stereo Vision for Moving Object (스테레오 비전 기반의 이동객체용 실시간 환경 인식 시스템)

  • Lee, Chung-Hee;Lim, Young-Chul;Kwon, Soon;Lee, Jong-Hun
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.3 no.3
    • /
    • pp.143-150
    • /
    • 2008
  • In this paper, we investigate a real-time environment recognition system based on stereo vision for moving object. This system consists of stereo matching, obstacle detection and distance estimation. In stereo matching part, depth maps can be obtained real road images captured adjustable baseline stereo vision system using belief propagation(BP) algorithm. In detection part, various obstacles are detected using only depth map in case of both v-disparity and column detection method under the real road environment. Finally in estimation part, asymmetric parabola fitting with NCC method improves estimation of obstacle detection. This stereo vision system can be applied to many applications such as unmanned vehicle and robot.

  • PDF

Vision-sensor-based Drivable Area Detection Technique for Environments with Changes in Road Elevation and Vegetation (도로의 높낮이 변화와 초목이 존재하는 환경에서의 비전 센서 기반)

  • Lee, Sangjae;Hyun, Jongkil;Kwon, Yeon Soo;Shim, Jae Hoon;Moon, Byungin
    • Journal of Sensor Science and Technology
    • /
    • v.28 no.2
    • /
    • pp.94-100
    • /
    • 2019
  • Drivable area detection is a major task in advanced driver assistance systems. For drivable area detection, several studies have proposed vision-sensor-based approaches. However, conventional drivable area detection methods that use vision sensors are not suitable for environments with changes in road elevation. In addition, if the boundary between the road and vegetation is not clear, judging a vegetation area as a drivable area becomes a problem. Therefore, this study proposes an accurate method of detecting drivable areas in environments in which road elevations change and vegetation exists. Experimental results show that when compared to the conventional method, the proposed method improves the average accuracy and recall of drivable area detection on the KITTI vision benchmark suite by 3.42%p and 8.37%p, respectively. In addition, when the proposed vegetation area removal method is applied, the average accuracy and recall are further improved by 6.43%p and 9.68%p, respectively.

Light Source Target Detection Algorithm for Vision-based UAV Recovery

  • Won, Dae-Yeon;Tahk, Min-Jea;Roh, Eun-Jung;Shin, Sung-Sik
    • International Journal of Aeronautical and Space Sciences
    • /
    • v.9 no.2
    • /
    • pp.114-120
    • /
    • 2008
  • In the vision-based recovery phase, a terminal guidance for the blended-wing UAV requires visual information of high accuracy. This paper presents the light source target design and detection algorithm for vision-based UAV recovery. We propose a recovery target design with red and green LEDs. This frame provides the relative position between the target and the UAV. The target detection algorithm includes HSV-based segmentation, morphology, and blob processing. These techniques are employed to give efficient detection results in day and night net recovery operations. The performance of the proposed target design and detection algorithm are evaluated through ground-based experiments.

Vision Chip for Edge and Motion Detection with a Function of Output Offset Cancellation (출력옵셋의 제거기능을 가지는 윤곽 및 움직임 검출용 시각칩)

  • Park, Jong-Ho;Kim, Jung-Hwan;Suh, Sung-Ho;Shin, Jang-Kyoo;Lee, Min-Ho
    • Journal of Sensor Science and Technology
    • /
    • v.13 no.3
    • /
    • pp.188-194
    • /
    • 2004
  • With a remarkable advance in CMOS (complimentary metal-oxide-semiconductor) process technology, a variety of vision sensors with signal processing circuits for complicated functions are actively being developed. Especially, as the principles of signal processing in human retina have been revealed, a series of vision chips imitating human retina have been reported. Human retina is able to detect the edge and motion of an object effectively. The edge detection among the several functions of the retina is accomplished by the cells called photoreceptor, horizontal cell and bipolar cell. We designed a CMOS vision chip by modeling cells of the retina as hardwares involved in edge and motion detection. The designed vision chip was fabricated using $0.6{\mu}m$ CMOS process and the characteristics were measured. Having reliable output characteristics, this chip can be used at the input stage for many applications, like targe tracking system, fingerprint recognition system, human-friendly robot system and etc.

Vision chip for edge detection with resolution improvement through simplification of unit-pixel circuit (단위 픽셀 회로의 간소화를 통해서 해상도를 향상시킨 이차원 윤곽 검출용 시각칩)

  • Sung, Dong-Kyu;Kong, Jae-Sung;Hyun, Hyo-Young;Shin, Jang-Kyoo
    • Journal of Sensor Science and Technology
    • /
    • v.17 no.1
    • /
    • pp.15-22
    • /
    • 2008
  • When designing image sensors including a CMOS vision chip for edge detection, resolution is a significant factor to evaluate the performance. It is hard to improve the resolution of a bio-inspired CMOS vision using a resistive network because the vision chip contains many circuits such as a resistive network and several signal processing circuits as well as photocircuits of general image sensors such as CMOS image sensor (CIS). Low resolution restricts the use of the application systems. In this paper, we improve the resolution through layout and circuit optimization. Furthermore, we have designed a printed circuit board using FPGA which controls the vision chip. The vision chip for edge detection has been designed and fabricated by using $0.35{\mu}m$ double-poly four-metal CMOS technology, and its output characteristics have been investigated.

STEREO VISION-BASED FORWARD OBSTACLE DETECTION

  • Jung, H.G.;Lee, Y.H.;Kim, B.J.;Yoon, P.J.;Kim, J.H.
    • International Journal of Automotive Technology
    • /
    • v.8 no.4
    • /
    • pp.493-504
    • /
    • 2007
  • This paper proposes a stereo vision-based forward obstacle detection and distance measurement method. In general, stereo vision-based obstacle detection methods in automotive applications can be classified into two categories: IPM (Inverse Perspective Mapping)-based and disparity histogram-based. The existing disparity histogram-based method was developed for stop-and-go applications. The proposed method extends the scope of the disparity histogram-based method to highway applications by 1) replacing the fixed rectangular ROI (Region Of Interest) with the traveling lane-based ROI, and 2) replacing the peak detection with a constant threshold with peak detection using the threshold-line and peakness evaluation. In order to increase the true positive rate while decreasing the false positive rate, multiple candidate peaks were generated and then verified by the edge feature correlation method. By testing the proposed method with images captured on the highway, it was shown that the proposed method was able to overcome problems in previous implementations while being applied successfully to highway collision warning/avoidance conditions, In addition, comparisons with laser radar showed that vision sensors with a wider FOV (Field Of View) provided faster responses to cutting-in vehicles. Finally, we integrated the proposed method into a longitudinal collision avoidance system. Experimental results showed that activated braking by risk assessment using the state of the ego-vehicle and measuring the distance to upcoming obstacles could successfully prevent collisions.

A Vision-based Damage Detection for Bridge Cables (교량케이블 영상기반 손상탐지)

  • Ho, Hoai-Nam;Lee, Jong-Jae
    • 한국방재학회:학술대회논문집
    • /
    • 2011.02a
    • /
    • pp.39-39
    • /
    • 2011
  • This study presents an effective vision-based system for cable bridge damage detection. In theory, cable bridges need to be inspected the outer as well as the inner part. Starting from August 2010, a new research project supported by Korea Ministry of Land, Transportation Maritime Affairs(MLTM) was initiated focusing on the damage detection of cable system. In this study, only the surface damage detection algorithm based on a vision-based system will be focused on, an overview of the vision-based cable damage detection is given in Fig. 1. Basically, the algorithm combines the image enhancement technique with principal component analysis(PCA) to detect damage on cable surfaces. In more detail, the input image from a camera is processed with image enhancement technique to improve image quality, and then it is projected into PCA sub-space. Finally, the Mahalanobis square distance is used for pattern recognition. The algorithm was verified through laboratory tests on three types of cable surface. The algorithm gave very good results, and the next step of this study is to implement the algorithm for real cable bridges.

  • PDF

A Survey on Vision Transformers for Object Detection Task (객체 탐지 과업에서의 트랜스포머 기반 모델의 특장점 분석 연구)

  • Jungmin, Ha;Hyunjong, Lee;Jungmin, Eom;Jaekoo, Lee
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.17 no.6
    • /
    • pp.319-327
    • /
    • 2022
  • Transformers are the most famous deep learning models that has achieved great success in natural language processing and also showed good performance on computer vision. In this survey, we categorized transformer-based models for computer vision, particularly object detection tasks and perform comprehensive comparative experiments to understand the characteristics of each model. Next, we evaluated the models subdivided into standard transformer, with key point attention, and adding attention with coordinates by performance comparison in terms of object detection accuracy and real-time performance. For performance comparison, we used two metrics: frame per second (FPS) and mean average precision (mAP). Finally, we confirmed the trends and relationships related to the detection and real-time performance of objects in several transformer models using various experiments.