• Title/Summary/Keyword: Computer vision technology

Search Result 666, Processing Time 0.026 seconds

Design and Implementation of the Stop line and Crosswalk Recognition Algorithm for Autonomous UGV (자율 주행 UGV를 위한 정지선과 횡단보도 인식 알고리즘 설계 및 구현)

  • Lee, Jae Hwan;Yoon, Heebyung
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.24 no.3
    • /
    • pp.271-278
    • /
    • 2014
  • In spite of that stop line and crosswalk should be aware of the most basic objects in transportation system, its features extracted are very limited. In addition to image-based recognition technology, laser and RF, GPS/INS recognition technology, it is difficult to recognize. For this reason, the limited research in this area has been done. In this paper, the algorithm to recognize the stop line and crosswalk is designed and implemented using image-based recognition technology with the images input through a vision sensor. This algorithm consists of three functions.; One is to select the area, in advance, needed for feature extraction in order to speed up the data processing, 'Region of Interest', another is to process the images only that white color is detected more than a certain proportion in order to remove the unnecessary operation, 'Color Pattern Inspection', the other is 'Feature Extraction and Recognition', which is to extract the edge features and compare this to the previously-modeled one to identify the stop line and crosswalk. For this, especially by using case based feature comparison algorithm, it can identify either both stop line and crosswalk exist or just one exists. Also the proposed algorithm is to develop existing researches by comparing and analysing effect of in-vehicle camera installation and changes in recognition rate of distance estimation and various constraints such as backlight and shadow.

Counting and Localizing Occupants using IR-UWB Radar and Machine Learning

  • Ji, Geonwoo;Lee, Changwon;Yun, Jaeseok
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.5
    • /
    • pp.1-9
    • /
    • 2022
  • Localization systems can be used with various circumstances like measuring population movement and rescue technology, even in security technology (like infiltration detection system). Vision sensors such as camera often used for localization is susceptible with light and temperature, and can cause invasion of privacy. In this paper, we used ultra-wideband radar technology (which is not limited by aforementioned problems) and machine learning techniques to measure the number and location of occupants in other indoor spaces behind the wall. We used four different algorithms and compared their results, including extremely randomized tree for four different situations; detect the number of occupants in a classroom, split the classroom into 28 locations and check the position of occupant, select one out of the 28 locations, divide it into 16 fine-grained locations, and check the position of occupant, and checking the positions of two occupants (existing in different locations). Overall, four algorithms showed good results and we verified that detecting the number and location of occupants are possible with high accuracy using machine learning. Also we have considered the possibility of service expansion using the oneM2M standard platform and expect to develop more service and products if this technology is used in various fields.

A modified U-net for crack segmentation by Self-Attention-Self-Adaption neuron and random elastic deformation

  • Zhao, Jin;Hu, Fangqiao;Qiao, Weidong;Zhai, Weida;Xu, Yang;Bao, Yuequan;Li, Hui
    • Smart Structures and Systems
    • /
    • v.29 no.1
    • /
    • pp.1-16
    • /
    • 2022
  • Despite recent breakthroughs in deep learning and computer vision fields, the pixel-wise identification of tiny objects in high-resolution images with complex disturbances remains challenging. This study proposes a modified U-net for tiny crack segmentation in real-world steel-box-girder bridges. The modified U-net adopts the common U-net framework and a novel Self-Attention-Self-Adaption (SASA) neuron as the fundamental computing element. The Self-Attention module applies softmax and gate operations to obtain the attention vector. It enables the neuron to focus on the most significant receptive fields when processing large-scale feature maps. The Self-Adaption module consists of a multiplayer perceptron subnet and achieves deeper feature extraction inside a single neuron. For data augmentation, a grid-based crack random elastic deformation (CRED) algorithm is designed to enrich the diversities and irregular shapes of distributed cracks. Grid-based uniform control nodes are first set on both input images and binary labels, random offsets are then employed on these control nodes, and bilinear interpolation is performed for the rest pixels. The proposed SASA neuron and CRED algorithm are simultaneously deployed to train the modified U-net. 200 raw images with a high resolution of 4928 × 3264 are collected, 160 for training and the rest 40 for the test. 512 × 512 patches are generated from the original images by a sliding window with an overlap of 256 as inputs. Results show that the average IoU between the recognized and ground-truth cracks reaches 0.409, which is 29.8% higher than the regular U-net. A five-fold cross-validation study is performed to verify that the proposed method is robust to different training and test images. Ablation experiments further demonstrate the effectiveness of the proposed SASA neuron and CRED algorithm. Promotions of the average IoU individually utilizing the SASA and CRED module add up to the final promotion of the full model, indicating that the SASA and CRED modules contribute to the different stages of model and data in the training process.

A Study on Swarm Robot-Based Invader-Enclosing Technique on Multiple Distributed Object Environments

  • Ko, Kwang-Eun;Park, Seung-Min;Park, Jun-Heong;Sim, Kwee-Bo
    • Journal of Electrical Engineering and Technology
    • /
    • v.6 no.6
    • /
    • pp.806-816
    • /
    • 2011
  • Interest about social security has recently increased in favor of safety for infrastructure. In addition, advances in computer vision and pattern recognition research are leading to video-based surveillance systems with improved scene analysis capabilities. However, such video surveillance systems, which are controlled by human operators, cannot actively cope with dynamic and anomalous events, such as having an invader in the corporate, commercial, or public sectors. For this reason, intelligent surveillance systems are increasingly needed to provide active social security services. In this study, we propose a core technique for intelligent surveillance system that is based on swarm robot technology. We present techniques for invader enclosing using swarm robots based on multiple distributed object environment. The proposed methods are composed of three main stages: location estimation of the object, specified object tracking, and decision of the cooperative behavior of the swarm robots. By using particle filter, object tracking and location estimation procedures are performed and a specified enclosing point for the swarm robots is located on the interactive positions in their coordinate system. Furthermore, the cooperative behaviors of the swarm robots are determined via the result of path navigation based on the combination of potential field and wall-following methods. The results of each stage are combined into the swarm robot-based invader-enclosing technique on multiple distributed object environments. Finally, several simulation results are provided to further discuss and verify the accuracy and effectiveness of the proposed techniques.

Deep Learning-based Action Recognition using Skeleton Joints Mapping (스켈레톤 조인트 매핑을 이용한 딥 러닝 기반 행동 인식)

  • Tasnim, Nusrat;Baek, Joong-Hwan
    • Journal of Advanced Navigation Technology
    • /
    • v.24 no.2
    • /
    • pp.155-162
    • /
    • 2020
  • Recently, with the development of computer vision and deep learning technology, research on human action recognition has been actively conducted for video analysis, video surveillance, interactive multimedia, and human machine interaction applications. Diverse techniques have been introduced for human action understanding and classification by many researchers using RGB image, depth image, skeleton and inertial data. However, skeleton-based action discrimination is still a challenging research topic for human machine-interaction. In this paper, we propose an end-to-end skeleton joints mapping of action for generating spatio-temporal image so-called dynamic image. Then, an efficient deep convolution neural network is devised to perform the classification among the action classes. We use publicly accessible UTD-MHAD skeleton dataset for evaluating the performance of the proposed method. As a result of the experiment, the proposed system shows better performance than the existing methods with high accuracy of 97.45%.

Error Correction Scheme in Location-based AR System Using Smartphone (스마트폰을 이용한 위치정보기반 AR 시스템에서의 부정합 현상 최소화를 위한 기법)

  • Lee, Ju-Yong;Kwon, Jun-Sik
    • Journal of Digital Contents Society
    • /
    • v.16 no.2
    • /
    • pp.179-187
    • /
    • 2015
  • Spread of smartphone creates various contents. Among many contents, AR application using Location Based Service(LBS) is needed widely. In this paper, we propose error correction algorithm for location-based Augmented Reality(AR) system using computer vision technology in android environment. This method that detects the early features with SURF(Speeded Up Robust Features) algorithm to minimize the mismatch and to reduce the operations, and tracks the detected, and applies it in mobile environment. We use the GPS data to retrieve the location information, and use the gyro sensor and G-sensor to get the pose estimation and direction information. However, the cumulative errors of location information cause the mismatch that and an object is not fixed, and we can not accept it the complete AR technology. Because AR needs many operations, implementation in mobile environment has many difficulties. The proposed approach minimizes the performance degradation in mobile environments, and are relatively simple to implement, and a variety of existing systems can be useful in a mobile environment.

Mobile Robot Control using Hand Shape Recognition (손 모양 인식을 이용한 모바일 로봇제어)

  • Kim, Young-Rae;Kim, Eun-Yi;Chang, Jae-Sik;Park, Se-Hyun
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.45 no.4
    • /
    • pp.34-40
    • /
    • 2008
  • This paper presents a vision based walking robot control system using hand shape recognition. To recognize hand shapes, the accurate hand boundary needs to be tracked in image obtained from moving camera. For this, we use an active contour model-based tracking approach with mean shift which reduces dependency of the active contour model to location of initial curve. The proposed system is composed of four modules: a hand detector, a hand tracker, a hand shape recognizer and a robot controller. The hand detector detects a skin color region, which has a specific shape, as hand in an image. Then, the hand tracking is performed using an active contour model with mean shift. Thereafter the hand shape recognition is performed using Hue moments. To assess the validity of the proposed system we tested the proposed system to a walking robot, RCB-1. The experimental results show the effectiveness of the proposed system.

OWC based Smart TV Remote Controller Design Using Flashlight

  • Mariappan, Vinayagam;Lee, Minwoo;Choi, Byunghoon;Kim, Jooseok;Lee, Jisung;Choi, Seongjhin
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.10 no.1
    • /
    • pp.71-76
    • /
    • 2018
  • The technology convergence of television, communication, and computing devices enables the rich social and entertaining experience through Smart TV in personal living space. The powerful smart TV computing platform allows to provide various user interaction interfaces like IR remote control, web based control, body gesture based control, etc. The presently used smart TV interaction user control methods are not efficient and user-friendly to access different type of media content and services and strongly required advanced way to control and access to the smart TV with easy user interface. This paper propose the optical wireless communication (OWC) based remote controller design for Smart TV using smart device Flashlights. In this approach, the user smart device act as a remote controller with touch based interactive smart device application and transfer the user control interface data to smart TV trough Flashlight using visible light communication method. The smart TV built-in camera follows the optical camera communication (OCC) principle to decode data and control smart TV user access functions according. This proposed method is not harmful as radio frequency (RF) radiation does it on human health and very simple to use as well user does need to any gesture moves to control the smart TV.

Digital Mirror using Particle System based on Motion Detection (움직임 감지 기반의 파티클 시스템을 이용한 디지털 거울)

  • Lim, Chan;Yun, Jae-Sun
    • The Journal of the Korea Contents Association
    • /
    • v.11 no.11
    • /
    • pp.62-69
    • /
    • 2011
  • Development of sensing technology and progress of digital media have been creating new art genre named interactive media art. digital mirror working based on convergence between computer vision technology and video art, is expressing reconstituted spectator's visual image through various mediums. From this aesthetical point and high accessibility towards spectators, many types of digital mirrors have been introducing. However, the majority of digital mirrors express visual images unrelated to degree of spectator's participation and this caused obstruction to spectator's continuous participation and interaction. This paper proposes digital mirror operated by spectator's movements read through particle system synchronized with motion detection algorithm based on analyzing image difference. This work extracted the data of spectator's movement by image processing and designed particle system changed by this data. And it expressed reconstructed spectator's image.

A Method for Eliminating Aiming Error of Unguided Anti-Tank Rocket Using Improved Target Tracking (향상된 표적 추적 기법을 이용한 무유도 대전차 로켓의 조준 오차 제거 방법)

  • Song, Jin-Mo;Kim, Tae-Wan;Park, Tai-Sun;Do, Joo-Cheol;Bae, Jong-sue
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.21 no.1
    • /
    • pp.47-60
    • /
    • 2018
  • In this paper, we proposed a method for eliminating aiming error of unguided anti-tank rocket using improved target tracking. Since predicted fire is necessary to hit moving targets with unguided rockets, a method was proposed to estimate the position and velocity of target using fire control system. However, such a method has a problem that the hit rate may be lowered due to the aiming error of the shooter. In order to solve this problem, we used an image-based target tracking method to correct error caused by the shooter. We also proposed a robust tracking method based on TLD(Tracking Learning Detection) considering characteristics of the FCS(Fire Control System) devices. To verify the performance of our proposed algorithm, we measured the target velocity using GPS and compared it with our estimation. It is proved that our method is robust to shooter's aiming error.