• Title/Summary/Keyword: Monocular Camera

Search Result 111, Processing Time 0.021 seconds

Robust Global Localization based on Environment map through Sensor Fusion (센서 융합을 통한 환경지도 기반의 강인한 전역 위치추정)

  • Jung, Min-Kuk;Song, Jae-Bok
    • The Journal of Korea Robotics Society
    • /
    • v.9 no.2
    • /
    • pp.96-103
    • /
    • 2014
  • Global localization is one of the essential issues for mobile robot navigation. In this study, an indoor global localization method is proposed which uses a Kinect sensor and a monocular upward-looking camera. The proposed method generates an environment map which consists of a grid map, a ceiling feature map from the upward-looking camera, and a spatial feature map obtained from the Kinect sensor. The method selects robot pose candidates using the spatial feature map and updates sample poses by particle filter based on the grid map. Localization success is determined by calculating the matching error from the ceiling feature map. In various experiments, the proposed method achieved a position accuracy of 0.12m and a position update speed of 10.4s, which is robust enough for real-world applications.

Ground Plane Detection Using Homography Matrix (호모그래피행렬을 이용한 노면검출)

  • Lee, Ki-Yong;Lee, Joon-Woong
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.17 no.10
    • /
    • pp.983-988
    • /
    • 2011
  • This paper presents a robust method for ground plane detection in vision-based applications based on a monocular sequence of images with a non-stationary camera. The proposed method, which is based on the reliable estimation of the homography between two frames taken from the sequence, aims at designing a practical system to detect road surface from traffic scenes. The homography is computed using a feature matching approach, which often gives rise to inaccurate matches or undesirable matches from out of the ground plane. Hence, the proposed homography estimation minimizes the effects from erroneous feature matching by the evaluation of the difference between the predicted and the observed matrices. The method is successfully demonstrated for the detection of road surface performed on experiments to fill an information void area taken place from geometric transformation applied to captured images by an in-vehicle camera system.

Pose Tracking of Moving Sensor using Monocular Camera and IMU Sensor

  • Jung, Sukwoo;Park, Seho;Lee, KyungTaek
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.8
    • /
    • pp.3011-3024
    • /
    • 2021
  • Pose estimation of the sensor is important issue in many applications such as robotics, navigation, tracking, and Augmented Reality. This paper proposes visual-inertial integration system appropriate for dynamically moving condition of the sensor. The orientation estimated from Inertial Measurement Unit (IMU) sensor is used to calculate the essential matrix based on the intrinsic parameters of the camera. Using the epipolar geometry, the outliers of the feature point matching are eliminated in the image sequences. The pose of the sensor can be obtained from the feature point matching. The use of IMU sensor can help initially eliminate erroneous point matches in the image of dynamic scene. After the outliers are removed from the feature points, these selected feature points matching relations are used to calculate the precise fundamental matrix. Finally, with the feature point matching relation, the pose of the sensor is estimated. The proposed procedure was implemented and tested, comparing with the existing methods. Experimental results have shown the effectiveness of the technique proposed in this paper.

Particle Filter Based Robust Multi-Human 3D Pose Estimation for Vehicle Safety Control (차량 안전 제어를 위한 파티클 필터 기반의 강건한 다중 인체 3차원 자세 추정)

  • Park, Joonsang;Park, Hyungwook
    • Journal of Auto-vehicle Safety Association
    • /
    • v.14 no.3
    • /
    • pp.71-76
    • /
    • 2022
  • In autonomous driving cars, 3D pose estimation can be one of the effective methods to enhance safety control for OOP (Out of Position) passengers. There have been many studies on human pose estimation using a camera. Previous methods, however, have limitations in automotive applications. Due to unexplainable failures, CNN methods are unreliable, and other methods perform poorly. This paper proposes robust real-time multi-human 3D pose estimation architecture in vehicle using monocular RGB camera. Using particle filter, our approach integrates CNN 2D/3D pose measurements with available information in vehicle. Computer simulations were performed to confirm the accuracy and robustness of the proposed algorithm.

Extended and Adaptive Inverse Perspective Mapping for Ground Representation of Autonomous Mobile Robot (모바일 자율 주행 로봇의 지면 표현을 위한 확장된 적응형 역투영 맵핑 방법)

  • Jooyong Park;Younggun Cho
    • The Journal of Korea Robotics Society
    • /
    • v.18 no.1
    • /
    • pp.59-65
    • /
    • 2023
  • This paper proposes an Extended and Adaptive Inverse Perspective Mapping (EA-IPM) model that can obtain an accurate bird's-eye view (BEV) from the forward-looking monocular camera on the sidewalk with various curves. While Inverse Perspective Mapping (IPM) is a good way to obtain ground information, conventional methods assume a fixed relationship between the camera and the ground. Due to the nature of the driving environment of the mobile robot, there are more walking environments with frequent motion changes than flat roads, which have a fatal effect on IPM results. Therefore, we have developed an extended IPM process to be applicable in IPM on sidewalks by adding a formula for complementary Y-derive processes and roll motions to the existing adaptive IPM model that is robust to pitch motions. To convince the performance of the proposed method, we evaluated our results on both synthetic and real road and sidewalk datasets.

Vision Based Estimation of 3-D Position of Target for Target Following Guidance/Control of UAV (무인 항공기의 목표물 추적을 위한 영상 기반 목표물 위치 추정)

  • Kim, Jong-Hun;Lee, Dae-Woo;Cho, Kyeum-Rae;Jo, Seon-Yeong;Kim, Jung-Ho;Han, Dong-In
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.14 no.12
    • /
    • pp.1205-1211
    • /
    • 2008
  • This paper describes methods to estimate 3-D position of target with respect to reference frame through monocular image from unmanned aerial vehicle (UAV). 3-D position of target is used as information for surveillance, recognition and attack. In this paper. 3-D position of target is estimated to make guidance and control law, which can follow target, user interested. It is necessary that position of target is measured in image to solve 3-D position of target. In this paper, kalman filter is used to track and output position of target in image. Estimation of target's 3-D position is possible using result of image tracking and information of UAV and camera. To estimate this, two algorithms are used. One is methode from arithmetic derivation of dynamics between UAV, carmer, and target. The other is LPV (Linear Parametric Varying). These methods have been run on simulation, and compared in this paper.

An integrated visual-inertial technique for structural displacement and velocity measurement

  • Chang, C.C.;Xiao, X.H.
    • Smart Structures and Systems
    • /
    • v.6 no.9
    • /
    • pp.1025-1039
    • /
    • 2010
  • Measuring displacement response for civil structures is very important for assessing their performance, safety and integrity. Recently, video-based techniques that utilize low-cost high-resolution digital cameras have been developed for such an application. These techniques however have relatively low sampling frequency and the results are usually contaminated with noises. In this study, an integrated visual-inertial measurement method that combines a monocular videogrammetric displacement measurement technique and a collocated accelerometer is proposed for displacement and velocity measurement of civil engineering structures. The monocular videogrammetric technique extracts three-dimensional translation and rotation of a planar target from an image sequence recorded by one camera. The obtained displacement is then fused with acceleration measured from a collocated accelerometer using a multi-rate Kalman filter with smoothing technique. This data fusion not only can improve the accuracy and the frequency bandwidth of displacement measurement but also provide estimate for velocity. The proposed measurement technique is illustrated by a shake table test and a pedestrian bridge test. Results show that the fusion of displacement and acceleration can mitigate their respective limitations and produce more accurate displacement and velocity responses with a broader frequency bandwidth.

Stereo Vision-based Visual Odometry Using Robust Visual Feature in Dynamic Environment (동적 환경에서 강인한 영상특징을 이용한 스테레오 비전 기반의 비주얼 오도메트리)

  • Jung, Sang-Jun;Song, Jae-Bok;Kang, Sin-Cheon
    • The Journal of Korea Robotics Society
    • /
    • v.3 no.4
    • /
    • pp.263-269
    • /
    • 2008
  • Visual odometry is a popular approach to estimating robot motion using a monocular or stereo camera. This paper proposes a novel visual odometry scheme using a stereo camera for robust estimation of a 6 DOF motion in the dynamic environment. The false results of feature matching and the uncertainty of depth information provided by the camera can generate the outliers which deteriorate the estimation. The outliers are removed by analyzing the magnitude histogram of the motion vector of the corresponding features and the RANSAC algorithm. The features extracted from a dynamic object such as a human also makes the motion estimation inaccurate. To eliminate the effect of a dynamic object, several candidates of dynamic objects are generated by clustering the 3D position of features and each candidate is checked based on the standard deviation of features on whether it is a real dynamic object or not. The accuracy and practicality of the proposed scheme are verified by several experiments and comparisons with both IMU and wheel-based odometry. It is shown that the proposed scheme works well when wheel slip occurs or dynamic objects exist.

  • PDF

Mobile Robot Navigation using Data Fusion Based on Camera and Ultrasonic Sensors Algorithm (카메라와 초음파센서 융합에 의한이동로봇의 주행 알고리즘)

  • Jang, Gi-Dong;Park, Sang-Keon;Han, Sung-Min;Lee, Kang-Woong
    • Journal of Advanced Navigation Technology
    • /
    • v.15 no.5
    • /
    • pp.696-704
    • /
    • 2011
  • In this paper, we propose a mobile robot navigation algorithm using data fusion of a monocular camera and ultrasonic sensors. Threshold values for binary image processing are generated by a fuzzy inference method using image data and data of ultrasonic sensors. Threshold value variations improve obstacle detection for mobile robot to move to the goal under poor illumination environments. Obstacles detected by data fusion of camera and ultrasonic sensors are expressed on the grid map and avoided using the circular planning algorithm. The performance of the proposed method is evaluated by experiments on the Pioneer 2-DX mobile robot in the indoor room with poor lights and a narrow corridor.

Mixing Collaborative and Hybrid Vision Devices for Robotic Applications (로봇 응용을 위한 협력 및 결합 비전 시스템)

  • Bazin, Jean-Charles;Kim, Sung-Heum;Choi, Dong-Geol;Lee, Joon-Young;Kweon, In-So
    • The Journal of Korea Robotics Society
    • /
    • v.6 no.3
    • /
    • pp.210-219
    • /
    • 2011
  • This paper studies how to combine devices such as monocular/stereo cameras, motors for panning/tilting, fisheye lens and convex mirrors, in order to solve vision-based robotic problems. To overcome the well-known trade-offs between optical properties, we present two mixed versions of the new systems. The first system is the robot photographer with a conventional pan/tilt perspective camera and fisheye lens. The second system is the omnidirectional detector for a complete 360-degree field-of-view surveillance system. We build an original device that combines a stereo-catadioptric camera and a pan/tilt stereo-perspective camera, and also apply it in the real environment. Compared to the previous systems, we show benefits of two proposed systems in aspects of maintaining both high-speed and high resolution with collaborative moving cameras and having enormous search space with hybrid configuration. The experimental results are provided to show the effectiveness of the mixing collaborative and hybrid systems.