• 제목/요약/키워드: visual odometry

검색결과 35건 처리시간 0.017초

무인 시스템의 자율 주행을 위한 영상기반 항법기술 동향 (Survey on Visual Navigation Technology for Unmanned Systems)

  • 김현진;서호성;김표진;이충근
    • 한국항행학회논문지
    • /
    • 제19권2호
    • /
    • pp.133-139
    • /
    • 2015
  • 이 논문에서는 영상정보를 기반으로 한 무인 시스템의 자율 항법기술에 대한 동향을 요약한다. 영상기반 항법기술로는 비주얼 서보잉, 비주얼 오도메트리, 영상 기반 SLAM(simultaneous localization and mapping)이 있다. 비주얼 서보잉은 목표 이미지와 현재 이미지 사이의 피쳐 차이로부터 원하는 속도 입력을 계산하여 무인 로봇을 목표 자세로 유도하는 데 사용된다. 비주얼 오도메트리는 무인 시스템이 영상정보를 바탕으로 자신의 이동 궤적을 추정하는 기술로, 기존의 dead-reckoning 방식보다 정확성을 향상시킬 수 있다. 영상 기반 SLAM은 무인 시스템이 영상 정보를 활용하여 미지의 환경에 대한 지도를 구축함과 동시에 자신의 위치를 결정해 나가는 기술로, 정확히 알지 못하는 환경에서 무인차량이나 무인기를 운용하는데 필수적이다. 이러한 기술들이 적용된 해외의 연구 사례들을 살펴봄으로써 영상기반 항법기술의 동향을 파악할 수 있었다.

1-Point Ransac Based Robust Visual Odometry

  • Nguyen, Van Cuong;Heo, Moon Beom;Jee, Gyu-In
    • Journal of Positioning, Navigation, and Timing
    • /
    • 제2권1호
    • /
    • pp.81-89
    • /
    • 2013
  • Many of the current visual odometry algorithms suffer from some extreme limitations such as requiring a high amount of computation time, complex algorithms, and not working in urban environments. In this paper, we present an approach that can solve all the above problems using a single camera. Using a planar motion assumption and Ackermann's principle of motion, we construct the vehicle's motion model as a circular planar motion (2DOF). Then, we adopt a 1-point method to improve the Ransac algorithm and the relative motion estimation. In the Ransac algorithm, we use a 1-point method to generate the hypothesis and then adopt the Levenberg-Marquardt method to minimize the geometric error function and verify inliers. In motion estimation, we combine the 1-point method with a simple least-square minimization solution to handle cases in which only a few feature points are present. The 1-point method is the key to speed up our visual odometry application to real-time systems. Finally, a Bundle Adjustment algorithm is adopted to refine the pose estimation. The results on real datasets in urban dynamic environments demonstrate the effectiveness of our proposed algorithm.

BIM model-based structural damage localization using visual-inertial odometry

  • Junyeon Chung;Kiyoung Kim;Hoon Sohn
    • Smart Structures and Systems
    • /
    • 제31권6호
    • /
    • pp.561-571
    • /
    • 2023
  • Ensuring the safety of a structure necessitates that repairs are carried out based on accurate inspections and records of damage information. Traditional methods of recording damage rely on individual paper-based documents, making it challenging for inspectors to accurately record damage locations and track chronological changes. Recent research has suggested the adoption of building information modeling (BIM) to record detailed damage information; however, localizing damages on a BIM model can be time-consuming. To overcome this limitation, this study proposes a method to automatically localize damages on a BIM model in real-time, utilizing consecutive images and measurements from an inertial measurement unit in close proximity to damages. The proposed method employs a visual-inertial odometry algorithm to estimate the camera pose, detect damages, and compute the damage location in the coordinate of a prebuilt BIM model. The feasibility and effectiveness of the proposed method were validated through an experiment conducted on a campus building. Results revealed that the proposed method successfully localized damages on the BIM model in real-time, with a root mean square error of 6.6 cm.

딥러닝 기반 영상 주행기록계와 단안 깊이 추정 및 기술을 위한 벤치마크 (Benchmark for Deep Learning based Visual Odometry and Monocular Depth Estimation)

  • 최혁두
    • 로봇학회논문지
    • /
    • 제14권2호
    • /
    • pp.114-121
    • /
    • 2019
  • This paper presents a new benchmark system for visual odometry (VO) and monocular depth estimation (MDE). As deep learning has become a key technology in computer vision, many researchers are trying to apply deep learning to VO and MDE. Just a couple of years ago, they were independently studied in a supervised way, but now they are coupled and trained together in an unsupervised way. However, before designing fancy models and losses, we have to customize datasets to use them for training and testing. After training, the model has to be compared with the existing models, which is also a huge burden. The benchmark provides input dataset ready-to-use for VO and MDE research in 'tfrecords' format and output dataset that includes model checkpoints and inference results of the existing models. It also provides various tools for data formatting, training, and evaluation. In the experiments, the exsiting models were evaluated to verify their performances presented in the corresponding papers and we found that the evaluation result is inferior to the presented performances.

Visual SLAM의 건설현장 실내 측위 활용성 분석 (Analysis of Applicability of Visual SLAM for Indoor Positioning in the Building Construction Site)

  • 김태진;박지원;이병민;배강민;윤세빈;김태훈
    • 한국건축시공학회:학술대회논문집
    • /
    • 한국건축시공학회 2022년도 가을 학술논문 발표대회
    • /
    • pp.47-48
    • /
    • 2022
  • The positioning technology that measures the position of a person or object is a key technology to deal with the location of the real coordinate system or converge the real and virtual worlds, such as digital twins, augmented reality, virtual reality, and autonomous driving. In estimating the location of a person or object at an indoor construction site, there are restrictions that it is impossible to receive location information from the outside, the communication infrastructure is insufficient, and it is difficult to install additional devices. Therefore, this study tested the direct sparse odometry algorithm, one of the visual Simultaneous Localization and Mapping (vSLAM) that estimate the current location and surrounding map using only image information, at an indoor construction site and analyzed its applicability as an indoor positioning technology. As a result, it was found that it is possible to properly estimate the surrounding map and the current location even in the indoor construction site, which has relatively few feature points. The results of this study can be used as reference data for researchers related to indoor positioning technology for construction sites in the future.

  • PDF

달 탐사 로버의 적응형 움직임 가중치에 따른 스테레오 준직접방식 비주얼 오도메트리 (Stereo Semi-direct Visual Odometry with Adaptive Motion Prior Weights of Lunar Exploration Rover)

  • 정재형;허세종;박찬국
    • 한국항공우주학회지
    • /
    • 제46권6호
    • /
    • pp.479-486
    • /
    • 2018
  • 위성항법시스템이 없는 달 표면에서 탐사 로버의 신뢰성 있는 항법성능을 확보하기 위해 관성측정장치나 카메라와 같은 추가적인 센서를 활용한 항법 알고리즘이 필수적이다. 일례로 미국의 화성 탐사 로버에 스테레오 카메라를 이용한 비주얼 오도메트리(VO)가 성공적으로 사용된 바 있다. 본 논문에서는 달 유사환경의 스테레오 흑백 이미지를 입력받아 달 탐사 로버의 6 자유도 움직임을 추정하였다. 제안하는 알고리즘은 희소 이미지 정렬 기반의 준직접방식 VO를 통해 연속된 이미지간의 상대 움직임을 추정한다. 또한 비선형성에 취약한 직접방식 VO를 보완하고자 최적화 시 로버의 움직임에 따른 가중치를 비용 함수에 고려하였고, 그 가중치는 이전 단계에서 계산된 포즈의 선형 함수로 제안한다. 본 논문에서 제안하는 로버의 움직임에 따른 가중치를 통해 실제 달 환경의 특성을 반영하는 토론토 대학의 달 유사환경 데이터셋에서 VO 성능이 향상됨을 확인하였다.

AprilTag and Stereo Visual Inertial Odometry (A-SVIO) based Mobile Assets Localization at Indoor Construction Sites

  • Khalid, Rabia;Khan, Muhammad;Anjum, Sharjeel;Park, Junsung;Lee, Doyeop;Park, Chansik
    • 국제학술발표논문집
    • /
    • The 9th International Conference on Construction Engineering and Project Management
    • /
    • pp.344-352
    • /
    • 2022
  • Accurate indoor localization of construction workers and mobile assets is essential in safety management. Existing positioning methods based on GPS, wireless, vision, or sensor based RTLS are erroneous or expensive in large-scale indoor environments. Tightly coupled sensor fusion mitigates these limitations. This research paper proposes a state-of-the-art positioning methodology, addressing the existing limitations, by integrating Stereo Visual Inertial Odometry (SVIO) with fiducial landmarks called AprilTags. SVIO determines the relative position of the moving assets or workers from the initial starting point. This relative position is transformed to an absolute position when AprilTag placed at various entry points is decoded. The proposed solution is tested on the NVIDIA ISAAC SIM virtual environment, where the trajectory of the indoor moving forklift is estimated. The results show accurate localization of the moving asset within any indoor or underground environment. The system can be utilized in various use cases to increase productivity and improve safety at construction sites, contributing towards 1) indoor monitoring of man machinery coactivity for collision avoidance and 2) precise real-time knowledge of who is doing what and where.

  • PDF

3D-2D 모션 추정을 위한 LiDAR 정보 보간 알고리즘 (LiDAR Data Interpolation Algorithm for 3D-2D Motion Estimation)

  • 전현호;고윤호
    • 한국멀티미디어학회논문지
    • /
    • 제20권12호
    • /
    • pp.1865-1873
    • /
    • 2017
  • The feature-based visual SLAM requires 3D positions for the extracted feature points to perform 3D-2D motion estimation. LiDAR can provide reliable and accurate 3D position information with low computational burden, while stereo camera has the problem of the impossibility of stereo matching in simple texture image region, the inaccuracy in depth value due to error contained in intrinsic and extrinsic camera parameter, and the limited number of depth value restricted by permissible stereo disparity. However, the sparsity of LiDAR data may increase the inaccuracy of motion estimation and can even lead to the result of motion estimation failure. Therefore, in this paper, we propose three interpolation methods which can be applied to interpolate sparse LiDAR data. Simulation results obtained by applying these three methods to a visual odometry algorithm demonstrates that the selective bilinear interpolation shows better performance in the view point of computation speed and accuracy.

2차원 라이다와 상업용 영상-관성 기반 주행 거리 기록계를 이용한 3차원 점 구름 지도 작성 시스템 개발 (Development of 3D Point Cloud Mapping System Using 2D LiDAR and Commercial Visual-inertial Odometry Sensor)

  • 문종식;이병윤
    • 대한임베디드공학회논문지
    • /
    • 제16권3호
    • /
    • pp.107-111
    • /
    • 2021
  • A 3D point cloud map is an essential elements in various fields, including precise autonomous navigation system. However, generating a 3D point cloud map using a single sensor has limitations due to the price of expensive sensor. In order to solve this problem, we propose a precise 3D mapping system using low-cost sensor fusion. Generating a point cloud map requires the process of estimating the current position and attitude, and describing the surrounding environment. In this paper, we utilized a commercial visual-inertial odometry sensor to estimate the current position and attitude states. Based on the state value, the 2D LiDAR measurement values describe the surrounding environment to create a point cloud map. To analyze the performance of the proposed algorithm, we compared the performance of the proposed algorithm and the 3D LiDAR-based SLAM (simultaneous localization and mapping) algorithm. As a result, it was confirmed that a precise 3D point cloud map can be generated with the low-cost sensor fusion system proposed in this paper.

다중 채널 동적 객체 정보 추정을 통한 특징점 기반 Visual SLAM (A New Feature-Based Visual SLAM Using Multi-Channel Dynamic Object Estimation)

  • 박근형;조형기
    • 대한임베디드공학회논문지
    • /
    • 제19권1호
    • /
    • pp.65-71
    • /
    • 2024
  • An indirect visual SLAM takes raw image data and exploits geometric information such as key-points and line edges. Due to various environmental changes, SLAM performance may decrease. The main problem is caused by dynamic objects especially in highly crowded environments. In this paper, we propose a robust feature-based visual SLAM, building on ORB-SLAM, via multi-channel dynamic objects estimation. An optical flow and deep learning-based object detection algorithm each estimate different types of dynamic object information. Proposed method incorporates two dynamic object information and creates multi-channel dynamic masks. In this method, information on actually moving dynamic objects and potential dynamic objects can be obtained. Finally, dynamic objects included in the masks are removed in feature extraction part. As a results, proposed method can obtain more precise camera poses. The superiority of our ORB-SLAM was verified to compared with conventional ORB-SLAM by the experiment using KITTI odometry dataset.