• Title/Summary/Keyword: 3-D localization

Search Result 357, Processing Time 0.03 seconds

Self-localization of a Mobile Robot for Decreasing the Error and VRML Image Overlay (오차 감소를 위한 이동로봇 Self-Localization과 VRML 영상오버레이 기법)

  • Kwon Bang-Hyun;Shon Eun-Ho;Kim Young-Chul;Chong Kil-To
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.12 no.4
    • /
    • pp.389-394
    • /
    • 2006
  • Inaccurate localization exposes a robot to many dangerous conditions. It could make a robot be moved to wrong direction or damaged by collision with surrounding obstacles. There are numerous approaches to self-localization, and there are different modalities as well (vision, laser range finders, ultrasonic sonars). Since sensor information is generally uncertain and contains noise, there are many researches to reduce the noise. But, the correctness is limited because most researches are based on statistical approach. The goal of our research is to measure more exact robot location by matching between built VRML 3D model and real vision image. To determine the position of mobile robot, landmark-localization technique has been applied. Landmarks are any detectable structure in the physical environment. Some use vertical lines, others use specially designed markers, In this paper, specially designed markers are used as landmarks. Given known focal length and a single image of three landmarks it is possible to compute the angular separation between the lines of sight of the landmarks. The image-processing and neural network pattern matching techniques are employed to recognize landmarks placed in a robot working environment. After self-localization, the 2D scene of the vision is overlaid with the VRML scene.

Localization Using 3D-Lidar Based Road Reflectivity Map and IPM Image (3D-Lidar 기반 도로 반사도 지도와 IPM 영상을 이용한 위치추정)

  • Jung, Tae-Ki;Song, Jong-Hwa;Im, Jun-Hyuck;Lee, Byung-Hyun;Jee, Gyu-In
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.22 no.12
    • /
    • pp.1061-1067
    • /
    • 2016
  • Position of the vehicle for driving is essential to autonomous navigation. However, there appears GPS position error due to multipath which is occurred by tall buildings in downtown area. In this paper, GPS position error is corrected by using camera sensor and highly accurate map made with 3D-Lidar. Input image through inverse perspective mapping is converted into top-view image, and it works out map matching with the map which has intensity of 3D-Lidar. Performance comparison was conducted between this method and traditional way which does map matching with input image after conversion of map to pinhole camera image. As a result, longitudinal error declined 49% and complexity declined 90%.

Object Detection and Localization on Map using Multiple Camera and Lidar Point Cloud

  • Pansipansi, Leonardo John;Jang, Minseok;Lee, Yonsik
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.422-424
    • /
    • 2021
  • In this paper, it leads the approach of fusing multiple RGB cameras for visual objects recognition based on deep learning with convolution neural network and 3D Light Detection and Ranging (LiDAR) to observe the environment and match into a 3D world in estimating the distance and position in a form of point cloud map. The goal of perception in multiple cameras are to extract the crucial static and dynamic objects around the autonomous vehicle, especially the blind spot which assists the AV to navigate according to the goal. Numerous cameras with object detection might tend slow-going the computer process in real-time. The computer vision convolution neural network algorithm to use for eradicating this problem use must suitable also to the capacity of the hardware. The localization of classified detected objects comes from the bases of a 3D point cloud environment. But first, the LiDAR point cloud data undergo parsing, and the used algorithm is based on the 3D Euclidean clustering method which gives an accurate on localizing the objects. We evaluated the method using our dataset that comes from VLP-16 and multiple cameras and the results show the completion of the method and multi-sensor fusion strategy.

  • PDF

VRML image overlay method for Robot's Self-Localization (VRML 영상오버레이기법을 이용한 로봇의 Self-Localization)

  • Sohn, Eun-Ho;Kwon, Bang-Hyun;Kim, Young-Chul;Chong, Kil-To
    • Proceedings of the KIEE Conference
    • /
    • 2006.04a
    • /
    • pp.318-320
    • /
    • 2006
  • Inaccurate localization exposes a robot to many dangerous conditions. It could make a robot be moved to wrong direction or damaged by collision with surrounding obstacles. There are numerous approaches to self-localization, and there are different modalities as well (vision, laser range finders, ultrasonic sonars). Since sensor information is generally uncertain and contains noise, there are many researches to reduce the noise. But, the correctness is limited because most researches are based on statistical approach. The goal of our research is to measure more exact robot location by matching between built VRML 3D model and real vision image. To determine the position of mobile robot, landmark-localitzation technique has been applied. Landmarks are any detectable structure in the physical environment. Some use vertical lines, others use specially designed markers, In this paper, specially designed markers are used as landmarks. Given known focal length and a single image of three landmarks it is possible to compute the angular separation between the lines of sight of the landmarks. The image-processing and neural network pattern matching techniques are employed to recognize landmarks placed in a robot working environment. After self-localization, the 2D scene of the vision is overlaid with the VRML scene.

  • PDF

Localization for Mobile Robot Using Vertical Lines

  • Kang, Chang-Hun;Ahn, Hyun-Sik
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2003.10a
    • /
    • pp.793-797
    • /
    • 2003
  • In this paper, we present a self-localization method for mobile robots using vertical line features of indoor environment. When a 2D map including feature points and color information is given, a mobile robot moves to the destination, and acquires images by one camera from the surroundings having vertical line edges. From the image, vertical line edges are detected, and pattern vectors meaning averaged color values of the left and right region of each line segment are computed. The pattern vectors are matched with the feature points of the map using the color information and the geometrical relationship of the points. From the perspective transformation of the corresponded points, nonlinear equations are derived. Localization is carried out from solving the equations by using Newton's method. Experimental results show that the proposed method using mono view is simple and applicable to indoor environment.

  • PDF

Visual Positioning System based on Voxel Labeling using Object Simultaneous Localization And Mapping

  • Jung, Tae-Won;Kim, In-Seon;Jung, Kye-Dong
    • International Journal of Advanced Culture Technology
    • /
    • v.9 no.4
    • /
    • pp.302-306
    • /
    • 2021
  • Indoor localization is one of the basic elements of Location-Based Service, such as indoor navigation, location-based precision marketing, spatial recognition of robotics, augmented reality, and mixed reality. We propose a Voxel Labeling-based visual positioning system using object simultaneous localization and mapping (SLAM). Our method is a method of determining a location through single image 3D cuboid object detection and object SLAM for indoor navigation, then mapping to create an indoor map, addressing it with voxels, and matching with a defined space. First, high-quality cuboids are created from sampling 2D bounding boxes and vanishing points for single image object detection. And after jointly optimizing the poses of cameras, objects, and points, it is a Visual Positioning System (VPS) through matching with the pose information of the object in the voxel database. Our method provided the spatial information needed to the user with improved location accuracy and direction estimation.

Vision-based Camera Localization using DEM and Mountain Image (DEM과 산영상을 이용한 비전기반 카메라 위치인식)

  • Cha Jeong-Hee
    • Journal of the Korea Society of Computer and Information
    • /
    • v.10 no.6 s.38
    • /
    • pp.177-186
    • /
    • 2005
  • In this Paper. we propose vision-based camera localization technique using 3D information which is created by mapping of DEM and mountain image. Typically, image features for localization have drawbacks, it is variable to camera viewpoint and after time information quantify increases . In this paper, we extract invariance features of geometry which is irrelevant to camera viewpoint and estimate camera extrinsic Parameter through accurate corresponding Points matching by Proposed similarity evaluation function and Graham search method we also propose 3D information creation method by using graphic theory and visual clues, The Proposed method has the three following stages; point features invariance vector extraction, 3D information creation, camera extrinsic Parameter estimation. In the experiments, we compare and analyse the proposed method with existing methods to demonstrate the superiority of the proposed methods.

  • PDF

The Underwater UUV Docking with 3D RF Signal Attenuation based Localization (UUV의 수중 도킹을 위한 전자기파 신호 기반의 위치인식 센서 개발)

  • Kwak, Kyungmin;Park, Daegil;Chung, Wan Kyun;Kim, Jinhyun
    • Journal of Sensor Science and Technology
    • /
    • v.26 no.3
    • /
    • pp.199-203
    • /
    • 2017
  • In this paper, we developed an underwater localization system for underwater robot docking using the electromagnetic wave attenuation model. Electromagnetic waves are generally known to be impossible to use in water environment. However, according to the conclusions of the previous studies on the attenuation characteristics in underwater, the attenuation pattern is uniform and its model was accurately proposed and verified in 3-dimensional space via the omnidirectional antenna. In this paper, a docking structure and localization sensor system are developed for a widely used cone type docking mechanism. First, we fabricated electromagnetic wave range sensor transmit modules. And a mobile sensor node is equipped with unmanned underwater vehicle(UUV)s. The mobile node senses the four different signal strength (RSS: Received Signal Strength) from fixed nodes, and the obtained RSS data are transformed to each distance information using the 3-Dimensional EM wave attenuation model. Then, the relative localization between the docking area and underwater robot can be achieved according to optimization algorithm. Finally, experimental results show the feasibility of the proposed localization system for the docking induction by comparing the errors in the actual position of the mobile node and the theoretical position through the model.

Development of Map Building Algorithm for Mobile Robot by Using RFID (모바일 로봇에서 RFID를 이용한 지도작성 알고리즘 개발)

  • Kim, Si-Seup;Seon, Jeong-An;Kee, Chang-Doo
    • Journal of the Korean Society of Manufacturing Technology Engineers
    • /
    • v.20 no.2
    • /
    • pp.133-138
    • /
    • 2011
  • RFID system can be used to improve object recognition, map building and localization for robot area. A novel method of indoor navigation system for a mobile robot is proposed using RFID technology. The mobile robot With a RFID reader and antenna is able to find what obstacles are located where in circumstance and can build the map similar to indoor circumstance by combining RFID information and distance data obtained from sensors. Using the map obtained, the mobile robot can avoid obstacles and finally reach the desired goal by $A^*$ algorithm. 3D map which has the advantage of robot navigation and manipulation is able to be built using z dimension of products. The proposed robot navigation system is proved to apply for SLAM and path planning in unknown circumstance through numerous experiments.