• Title/Summary/Keyword: SLAM Technology

Search Result 83, Processing Time 0.028 seconds

Development of Drone Racing Simulator using SLAM Technology and Reconstruction of Simulated Environments (SLAM 기술을 활용한 가상 환경 복원 및 드론 레이싱 시뮬레이션 제작)

  • Park, Yonghee;Yu, Seunghyun;Lee, Jaegwang;Jeong, Jonghyeon;Jo, Junhyeong;Kim, Soyeon;Oh, Hyejun;Moon, Hyungpil
    • The Journal of Korea Robotics Society
    • /
    • v.16 no.3
    • /
    • pp.245-249
    • /
    • 2021
  • In this paper, we present novel simulation contents for drone racing and autonomous flight of drone. With Depth camera and SLAM, we conducted mapping 3 dimensional environment through RTAB-map. The 3 dimensional map is represented by point cloud data. After that we recovered this data in Unreal Engine. This recovered raw data reflects real data that includes noise and outlier. Also we built drone racing contents like gate and obstacles for evaluating drone flight in Unreal Engine. Then we implemented both HITL and SITL by using AirSim which offers flight controller and ROS api. Finally we show autonomous flight of drone with ROS and AirSim. Drone can fly in real place and sensor property so drone experiences real flight even in the simulation world. Our simulation framework increases practicality than other common simulation that ignore real environment and sensor.

Development of autonomous driving route guidance robot using SLAM technology (SLAM 기술을 이용한 자율주행 경로 안내 로봇 개발)

  • Seung, Sang-jun;Lee, Ji-hwan;Jo, Min-je;Shin, Chun-ho;Kim, Do-yeon;Park, Yang-woo
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2021.01a
    • /
    • pp.153-154
    • /
    • 2021
  • 본 논문에서는 ROS(Robot Operating System)를 기반으로 한 로봇(Robot)에 LiDAR 센서를 설치하여 SLAM(Simultaneous Localization and Mapping) 기술인 동시적 위치 추적 지도 작성 기법을 이용하여 실내 맵 정보를 습득하고, 이를 기반으로 장애물과 건물 실내를 안전하고 정확하게 이동할 수 있도록 하였다. 또한 로봇에 자바에서 제공하는 개발 툴킷 Swing 및 AWT 라이브러리를 이용하여 GUI(Graphical User Interface)를 구현하였고 터치스크린을 장착하여 사용자가 원하는 제품을 선택하고 선택한 제품의 목적지를 습득한 맵을 토대로 좌표 값을 설정하여 ROS에서 지원하는 이동 프로세스를 실행시켜 목적지까지 경로를 설정하고 자율 주행하게 된다.

  • PDF

Development of Low Cost Autonomous-Driving Delivery Robot System Using SLAM Technology (SLAM 기술을 활용한 저가형 자율주행 배달 로봇 시스템 개발)

  • Donghoon Lee;Jehyun Park;Kyunghoon Jung
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.5
    • /
    • pp.249-257
    • /
    • 2023
  • This paper discusses the increasing need for autonomous delivery robots due to the current growth in the delivery market, rising delivery fees, high costs of hiring delivery personnel, and the need for contactless services. Additionally, the cost of hardware and complex software systems required to build and operate autonomous delivery robots is high. To provide a low-cost alternative to this, this paper proposes a autonomous delivery robot platform using a low-cost sensor combination of 2D LIDAR, depth camera and tracking camera to replace the existing expensive 3D LIDAR. The proposed robot was developed using the RTAB-Map SLAM open source package for 2D mapping and overcomes the limitations of low-cost sensors by using the convex hull algorithm. The paper details the hardware and software configuration of the robot and presents the results of driving experiments. The proposed platform has significant potential for various industries, including the delivery and other industries.

Collision Avoidance Using Omni Vision SLAM Based on Fisheye Image (어안 이미지 기반의 전방향 영상 SLAM을 이용한 충돌 회피)

  • Choi, Yun Won;Choi, Jeong Won;Im, Sung Gyu;Lee, Suk Gyu
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.22 no.3
    • /
    • pp.210-216
    • /
    • 2016
  • This paper presents a novel collision avoidance technique for mobile robots based on omni-directional vision simultaneous localization and mapping (SLAM). This method estimates the avoidance path and speed of a robot from the location of an obstacle, which can be detected using the Lucas-Kanade Optical Flow in images obtained through fish-eye cameras mounted on the robots. The conventional methods suggest avoidance paths by constructing an arbitrary force field around the obstacle found in the complete map obtained through the SLAM. Robots can also avoid obstacles by using the speed command based on the robot modeling and curved movement path of the robot. The recent research has been improved by optimizing the algorithm for the actual robot. However, research related to a robot using omni-directional vision SLAM to acquire around information at once has been comparatively less studied. The robot with the proposed algorithm avoids obstacles according to the estimated avoidance path based on the map obtained through an omni-directional vision SLAM using a fisheye image, and returns to the original path. In particular, it avoids the obstacles with various speed and direction using acceleration components based on motion information obtained by analyzing around the obstacles. The experimental results confirm the reliability of an avoidance algorithm through comparison between position obtained by the proposed algorithm and the real position collected while avoiding the obstacles.

SLAM : An Efficient Buffer Management Strategy using Spatial Locality of Spatial Data (SLAM : 공간 데이타의 공간적 근접성을 이용한 효율적인 버퍼관리기법)

  • An, Jae-Yong;Min, Jun-Gi;Jeong, Jin-Wan
    • Journal of KIISE:Databases
    • /
    • v.29 no.5
    • /
    • pp.393-403
    • /
    • 2002
  • One of the major issues of DBMS is the buffer management. Because fetching data from the database disk is costly, the number of disk I/O's must be minimized in order to improve the DBMS performance. Although there have been many buffer management strategies to minimize the disk I/O, those strategies usually focused on just the temporal locality. Since there are the spatial locality as well as the temporal locality in the spatial database, strategies using only the temporal locality cannot achieve the optimal performance in the spatial database. In this paper, we propose a new buffer management strategy, the Spatial Locality Area Measure(SLAM) strategy, that considers not only the temporal locality but also the spatial locality. The SLAM buffer management strategy consists of two core structures, the SLM-tree and the M-LRU. We show the efficiency of the proposed strategy through experiments over various buffer sizes and reference frequencies.

Unmanned Aerial Vehicle Recovery Using a Simultaneous Localization and Mapping Algorithm without the Aid of Global Positioning System

  • Lee, Chang-Hun;Tahk, Min-Jea
    • International Journal of Aeronautical and Space Sciences
    • /
    • v.11 no.2
    • /
    • pp.98-109
    • /
    • 2010
  • This paper deals with a new method of unmanned aerial vehicle (UAV) recovery when a UAV fails to get a global positioning system (GPS) signal at an unprepared site. The proposed method is based on the simultaneous localization and mapping (SLAM) algorithm. It is a process by which a vehicle can build a map of an unknown environment and simultaneously use this map to determine its position. Extensive research on SLAM algorithms proves that the error in the map reaches a lower limit, which is a function of the error that existed when the first observation was made. For this reason, the proposed method can help an inertial navigation system to prevent its error of divergence with regard to the vehicle position. In other words, it is possible that a UAV can navigate with reasonable positional accuracy in an unknown environment without the aid of GPS. This is the main idea of the present paper. Especially, this paper focuses on path planning that maximizes the discussed ability of a SLAM algorithm. In this work, a SLAM algorithm based on extended Kalman filter is used. For simplicity's sake, a blimp-type of UAV model is discussed and three-dimensional pointed-shape landmarks are considered. Finally, the proposed method is evaluated by a number of simulations.

3D Simultaneous Localization and Map Building (SLAM) using a 2D Laser Range Finder based on Vertical/Horizontal Planar Polygons (2차원 레이저 거리계를 이용한 수직/수평 다각평면 기반의 위치인식 및 3차원 지도제작)

  • Lee, Seungeun;Kim, Byung-Kook
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.20 no.11
    • /
    • pp.1153-1163
    • /
    • 2014
  • An efficient 3D SLAM (Simultaneous Localization and Map Building) method is developed for urban building environments using a tilted 2D LRF (Laser Range Finder), in which a 3D map is composed of perpendicular/horizontal planar polygons. While the mobile robot is moving, from the LRF scan distance data in each scan period, line segments on the scan plane are successively extracted. We propose an "expected line segment" concept for matching: to add each of these scan line segments to the most suitable line segment group for each perpendicular/horizontal planar polygon in the 3D map. After performing 2D localization to determine the pose of the mobile robot, we construct updated perpendicular/horizontal infinite planes and then determine their boundaries to obtain the perpendicular/horizontal planar polygons which constitute our 3D map. Finally, the proposed SLAM algorithm is validated via extensive simulations and experiments.

Vision-based Localization for AUVs using Weighted Template Matching in a Structured Environment (구조화된 환경에서의 가중치 템플릿 매칭을 이용한 자율 수중 로봇의 비전 기반 위치 인식)

  • Kim, Donghoon;Lee, Donghwa;Myung, Hyun;Choi, Hyun-Taek
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.19 no.8
    • /
    • pp.667-675
    • /
    • 2013
  • This paper presents vision-based techniques for underwater landmark detection, map-based localization, and SLAM (Simultaneous Localization and Mapping) in structured underwater environments. A variety of underwater tasks require an underwater robot to be able to successfully perform autonomous navigation, but the available sensors for accurate localization are limited. A vision sensor among the available sensors is very useful for performing short range tasks, in spite of harsh underwater conditions including low visibility, noise, and large areas of featureless topography. To overcome these problems and to a utilize vision sensor for underwater localization, we propose a novel vision-based object detection technique to be applied to MCL (Monte Carlo Localization) and EKF (Extended Kalman Filter)-based SLAM algorithms. In the image processing step, a weighted correlation coefficient-based template matching and color-based image segmentation method are proposed to improve the conventional approach. In the localization step, in order to apply the landmark detection results to MCL and EKF-SLAM, dead-reckoning information and landmark detection results are used for prediction and update phases, respectively. The performance of the proposed technique is evaluated by experiments with an underwater robot platform in an indoor water tank and the results are discussed.

Survey on Visual Navigation Technology for Unmanned Systems (무인 시스템의 자율 주행을 위한 영상기반 항법기술 동향)

  • Kim, Hyoun-Jin;Seo, Hoseong;Kim, Pyojin;Lee, Chung-Keun
    • Journal of Advanced Navigation Technology
    • /
    • v.19 no.2
    • /
    • pp.133-139
    • /
    • 2015
  • This paper surveys vision based autonomous navigation technologies for unmanned systems. Main branches of visual navigation technologies are visual servoing, visual odometry, and visual simultaneous localization and mapping (SLAM). Visual servoing provides velocity input which guides mobile system to desired pose. This input velocity is calculated from feature difference between desired image and acquired image. Visual odometry is the technology that estimates the relative pose between frames of consecutive image. This can improve the accuracy when compared with the exisiting dead-reckoning methods. Visual SLAM aims for constructing map of unknown environment and determining mobile system's location simultaneously, which is essential for operation of unmanned systems in unknown environments. The trend of visual navigation is grasped by examining foreign research cases related to visual navigation technology.

Augmented Feature Point Initialization Method for Vision/Lidar Aided 6-DoF Bearing-Only Inertial SLAM

  • Yun, Sukchang;Lee, Byoungjin;Kim, Yeon-Jo;Lee, Young Jae;Sung, Sangkyung
    • Journal of Electrical Engineering and Technology
    • /
    • v.11 no.6
    • /
    • pp.1846-1856
    • /
    • 2016
  • This study proposes a novel feature point initialization method in order to improve the accuracy of feature point positions by fusing a vision sensor and a lidar. The initialization is a process that determines three dimensional positions of feature points through two dimensional image data, which has a direct influence on performance of a 6-DoF bearing-only SLAM. Prior to the initialization, an extrinsic calibration method which estimates rotational and translational relationships between a vision sensor and lidar using multiple calibration tools was employed, then the feature point initialization method based on the estimated extrinsic calibration parameters was presented. In this process, in order to improve performance of the accuracy of the initialized feature points, an iterative automatic scaling parameter tuning technique was presented. The validity of the proposed feature point initialization method was verified in a 6-DoF bearing-only SLAM framework through an indoor and outdoor tests that compare estimation performance with the previous initialization method.