• Title/Summary/Keyword: Vision based tracking

Search Result 405, Processing Time 0.026 seconds

Automatic Detection of Dissimilar Regions through Multiple Feature Analysis (다중의 특징 분석을 통한 비 유사 영역의 자동적인 검출)

  • Jang, Seok-Woo;Jung, Myunghee
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.2
    • /
    • pp.160-166
    • /
    • 2020
  • As mobile-based hardware technology develops, many kinds of applications are also being developed. In addition, there is an increasing demand to automatically check that the interface of these applications works correctly. In this paper, we describe a method for accurately detecting faulty images from applications by comparing major characteristics from input color images. For this purpose, our method first extracts major characteristics of the input image, then calculates the differences in the extracted major features, and decides if the test image is a normal image or a faulty image dissimilar to the reference image. Experiment results show that the suggested approach robustly determines similar and dissimilar images by comparing major characteristics from input color images. The suggested method is expected to be useful in many real application areas related to computer vision, like video indexing, object detection and tracking, image surveillance, and so on.

A Study on Hand Gesture Recognition with Low-Resolution Hand Images (저해상도 손 제스처 영상 인식에 대한 연구)

  • Ahn, Jung-Ho
    • Journal of Satellite, Information and Communications
    • /
    • v.9 no.1
    • /
    • pp.57-64
    • /
    • 2014
  • Recently, many human-friendly communication methods have been studied for human-machine interface(HMI) without using any physical devices. One of them is the vision-based gesture recognition that this paper deals with. In this paper, we define some gestures for interaction with objects in a predefined virtual world, and propose an efficient method to recognize them. For preprocessing, we detect and track the both hands, and extract their silhouettes from the low-resolution hand images captured by a webcam. We modeled skin color by two Gaussian distributions in RGB color space and use blob-matching method to detect and track the hands. Applying the foodfill algorithm we extracted hand silhouettes and recognize the hand shapes of Thumb-Up, Palm and Cross by detecting and analyzing their modes. Then, with analyzing the context of hand movement, we recognized five predefined one-hand or both-hand gestures. Assuming that one main user shows up for accurate hand detection, the proposed gesture recognition method has been proved its efficiency and accuracy in many real-time demos.

A Study on Real Time Gaze Discrimination System using GRNN (GRNN을 이용한 실시간 시선 식별 시스템에 관한 연구)

  • Lee Young-Sik;Bae Cheol-Soo
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.9 no.2
    • /
    • pp.322-329
    • /
    • 2005
  • This paper describes a computer vision system based on active IR illumination for real-time gaze discrimination system. Unlike most of the existing gaze discrimination techniques, which often require assuming a static head to work well and require a cumbersome calibration process for each person, our gaze discrimination system can perform robust and accurate gaze estimation without calibration and under rather significant head movement. This is made possible by a new gaze calibration procedure that identifies the mapping from pupil parameters to screen coordinates using generalized regression neural networks (GRNNS). With GRNNS, the mapping does not have to be an analytical function and head movement is explicitly accounted for by the gaze mapping function. furthermore, the mapping function can generalize to other individuals not used in the training. To further improve the gaze estimation accuracy, we employ a reclassification scheme that deals with the classes that tend to be misclassified. This leads to a 10$\%$ improvement in classification error. The angular gaze accuracy is about $5^{circ}$horizontally and $8^{circ}$vertically. The effectiveness of our gaze tracker is demonstrated by experiments that involve gaze-contingent interactive graphic display.

Automatic Person Identification using Multiple Cues

  • Swangpol, Danuwat;Chalidabhongse, Thanarat
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.1202-1205
    • /
    • 2005
  • This paper describes a method for vision-based person identification that can detect, track, and recognize person from video using multiple cues: height and dressing colors. The method does not require constrained target's pose or fully frontal face image to identify the person. First, the system, which is connected to a pan-tilt-zoom camera, detects target using motion detection and human cardboard model. The system keeps tracking the moving target while it is trying to identify whether it is a human and identify who it is among the registered persons in the database. To segment the moving target from the background scene, we employ a version of background subtraction technique and some spatial filtering. Once the target is segmented, we then align the target with the generic human cardboard model to verify whether the detected target is a human. If the target is identified as a human, the card board model is also used to segment the body parts to obtain some salient features such as head, torso, and legs. The whole body silhouette is also analyzed to obtain the target's shape information such as height and slimness. We then use these multiple cues (at present, we uses shirt color, trousers color, and body height) to recognize the target using a supervised self-organization process. We preliminary tested the system on a set of 5 subjects with multiple clothes. The recognition rate is 100% if the person is wearing the clothes that were learned before. In case a person wears new dresses the system fail to identify. This means height is not enough to classify persons. We plan to extend the work by adding more cues such as skin color, and face recognition by utilizing the zoom capability of the camera to obtain high resolution view of face; then, evaluate the system with more subjects.

  • PDF

A Study on Motion Detection of Object Using Active Block Matching Algorithm (능동적 블록정합기법을 이용한 객체의 움직임 검출에 관한 연구)

  • Lee Chang-Soo;Park Mi-Og;Lee Kyung-Seok
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.31 no.4C
    • /
    • pp.407-416
    • /
    • 2006
  • It is difficult for the movement detection of an object through a camera to detect exact movement because of unnecessary noises and changes of the light. It can be recognized as a background, when there is no movement after the inflow of an object. Therefore, It is necessary to fast search algorithm for tracking and extract of object that is realtime image. In this thesis, we evaluate the difference of the input vision based on initial image and replace some pixels in process of time. When there is a big difference between background image and input image, we decide it is the point of the time of the object input and then extract boundary point of it. The extracted boundary point detects precise movement of the object by creating minimum block of it and searching block that maintaining distance. The designed and embodied system shows more than 95% accuracy in the performance test.

Conceptual Study on Coaxial Rotorcraft UAV for teaming operation with UGV (무인지상차량과의 합동운용을 위한 동축반전 회전익형 무인항공기 개념연구)

  • Byun, Young-Seop;Song, Jun-Beom;Song, Woo-Jin;Kim, Jeong;Kang, Beom-Soo
    • Journal of the Korean Society for Aeronautical & Space Sciences
    • /
    • v.39 no.5
    • /
    • pp.458-465
    • /
    • 2011
  • UAV-UGV teaming concept has been proposed that can compensate for weak points of each platform by providing carrying, launching, recovery and recharging capability for the VTOL-UAV through the host UGV. The teaming concept can expand the observation envelop of the UGV and extend the operational capability of the UAV through mechanical combination of each system. The spherical-shaped coaxial rotorcraft UAV is suggested to provide flexible and precise interface between two systems. Hybrid navigation solution that included vision-based target tracking method for precision landing is investigated and its experimental study is performed. Feasibility study on length-variable rotor to provide the compact configuration of the loaded rotorcraft platform is also described.

A Study of the Next Generation Ubiquitous Flight Information System and a Design (차세대 유비쿼터스 비행정보 시스템의 연구와 설계)

  • Park, Wan-Soon;Yang, Hae-Sool
    • Journal of the Korea Society of Computer and Information
    • /
    • v.13 no.4
    • /
    • pp.221-230
    • /
    • 2008
  • We Studied current Flight Information System everyone employing, and this paper designed a next generation Ubiquitous Flight Information System. Studied a Ubiquitous Flight Information System regarding AVOD, game TV, cellular phone, cabin Internet use for a customer, and airline and control of Air Traffic Service, Air Traffic Control, in seat reservation, transportation, ticketing and luggage back-tracking and airplanes in earthly service. We Designed for a Next Generation Ubiquitous Flight Information System and for necessary for network security, system security of information security system, and derived from comparative analysis of improvement point and difference to existing Ubiquitous Flight Information System and Next Generation Ubiquitous Flight Information as was based on result of research. Present vision to aerial the result of research world of this paper related industry, and ensure safety and communication efficiency and a customer and flight satisfaction, efficiency, and will reclaim the new horizon to the Flight Information Industry.

  • PDF

A method of improving the quality of 3D images acquired from RGB-depth camera (깊이 영상 카메라로부터 획득된 3D 영상의 품질 향상 방법)

  • Park, Byung-Seo;Kim, Dong-Wook;Seo, Young-Ho
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.5
    • /
    • pp.637-644
    • /
    • 2021
  • In general, in the fields of computer vision, robotics, and augmented reality, the importance of 3D space and 3D object detection and recognition technology has emerged. In particular, since it is possible to acquire RGB images and depth images in real time through an image sensor using Microsoft Kinect method, many changes have been made to object detection, tracking and recognition studies. In this paper, we propose a method to improve the quality of 3D reconstructed images by processing images acquired through a depth-based (RGB-Depth) camera on a multi-view camera system. In this paper, a method of removing noise outside an object by applying a mask acquired from a color image and a method of applying a combined filtering operation to obtain the difference in depth information between pixels inside the object is proposed. Through each experiment result, it was confirmed that the proposed method can effectively remove noise and improve the quality of 3D reconstructed image.

Extraction of Skin Regions through Filtering-based Noise Removal (필터링 기반의 잡음 제거를 통한 피부 영역의 추출)

  • Jang, Seok-Woo
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.12
    • /
    • pp.672-678
    • /
    • 2020
  • Ultra-high-speed images that accurately depict the minute movements of objects have become common as low-cost and high-performance cameras that can film at high speeds have emerged. In this paper, the proposed method removes unexpected noise contained in images after input at high speed, and then extracts an area of interest that can represent personal information, such as skin areas, from the image in which noise has been removed. In this paper, noise generated by abnormal electrical signals is removed by applying bilateral filters. A color model created through pre-learning is then used to extract the area of interest that represents the personal information contained within the image. Experimental results show that the introduced algorithms remove noise from high-speed images and then extract the area of interest robustly. The approach presented in this paper is expected to be useful in various applications related to computer vision, such as image preprocessing, noise elimination, tracking and monitoring of target areas, etc.

Analysis System for Public Interest Report Video of Traffic Law Violation based on Deep Learning Algorithms (딥러닝 알고리즘 기반 교통법규 위반 공익신고 영상 분석 시스템)

  • Min-Seong Choi;Mi-Kyeong Moon
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.18 no.1
    • /
    • pp.63-70
    • /
    • 2023
  • Due to the spread of high-definition black boxes and the introduction of mobile applications such as 'Smart Citizens Report' and 'Safety Report', the number of public interest reports for violations of Traffic Law has increased rapidly, resulting in shortage of police personnel to handle them. In this paper, we describe the development of a system that can automatically detect lane violations which account for the largest proportion of public interest reporting videos for violations of traffic laws, using deep learning algorithms. In this study, a method for recognizing a vehicle and a solid line object using a YOLO model and a Lanenet model, a method for tracking an object individually using a deep sort algorithm, and a method for detecting lane change violations by recognizing the overlapping range of a vehicle object's bounding box and a solid line object are described. Using this system, it is expected that the shortage of police personnel in charge will be resolved.