• Title/Summary/Keyword: Stereo Image Matching

Search Result 413, Processing Time 0.025 seconds

Range finding algorithm of equidistance stereo catadioptric mirror (등거리 스테레오 전방위 렌즈 영상에 대한 위치 측정 알고리즘)

  • Choi, Young-Ho
    • Journal of Internet Computing and Services
    • /
    • v.6 no.6
    • /
    • pp.149-161
    • /
    • 2005
  • Catadioptric mirrors are widely used in automatic surveillance system. The major drawback of catadioptric mirror is its unequal image resolution. Equidistance catadioptric mirror can be the solution to this problem. Even double panoramic structure can generate stereo images with single camera system. So two images obtained from double panoramic equidistance catadioptric mirror can be used in finding the depth and height values of object's points. But compared to the single catadioptric mirror. the image size of double panoramic system is relatively small. This leads to the severe accuracy problem in estimation. The exact axial alignment and the exact mount of mirror are the sources that can be avoided but the focal length variation is inevitable. In this paper, the effects of focal length variation on the computation of depth and height of object' point are explained and the effective focal length finding algorithm, using the assumption that the object's viewing angles are almost same in stereo images, is presented.

  • PDF

An Accurate Moving Distance Measurement Using the Rear-View Images in Parking Assistant Systems (후방영상 기반 주차 보조 시스템에서 정밀 이동거리 추출 기법)

  • Kim, Ho-Young;Lee, Seong-Won
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37C no.12
    • /
    • pp.1271-1280
    • /
    • 2012
  • In the recent parking assistant systems, finding out the distance to the object behind a car is often performed by the range sensors such as ultrasonic sensors, radars. However, the installation of additional sensors on the used vehicle could be difficult and require extra cost. On the other hand, the motion stereo technique that extracts distance information using only an image sensor was also proposed. However, In the stereo rectification step, the motion stereo requires good features and exacts matching result. In this paper, we propose a fast algorithm that extracts the accurate distance information for the parallel parking situation using the consecutive images that is acquired by a rear-view camera. The proposed algorithm uses the quadrangle transform of the image, the horizontal line integral projection, and the blocking-based correlation measurement. In the experiment with the magna parallel test sequence, the result shows that the line-accurate distance measurement with the image sequence from the rear-view camera is possible.

A 3-D Vision Sensor Implementation on Multiple DSPs TMS320C31 (다중 TMS320C31 DSP를 사용한 3-D 비젼센서 Implementation)

  • Oksenhendler, V.;Bensrhair, Abdelaziz;Miche, Pierre;Lee, Sang-Goog
    • Journal of Sensor Science and Technology
    • /
    • v.7 no.2
    • /
    • pp.124-130
    • /
    • 1998
  • High-speed 3D vision systems are essential for autonomous robot or vehicle control applications. In our study, a stereo vision process has been developed. It consists of three steps : extraction of edges in right and left images, matching corresponding edges and calculation of the 3D map. This process is implemented in a VME 150/40 Imaging Technology vision system. It is a modular system composed by a display, an acquisition, a four Mbytes image frame memory, and three computational cards. Programmable accelerator computational modules are running at 40 MHz and are based on TMS320C31 DSP with a $64{\times}32$ bit instruction cache and two $1024{\times}32$ bit internal RAMs. Each is equipped with 512 Kbytes static RAM, 4 Mbytes image memory, 1 Mbytes flash EEPROM and a serial port. Data transfers and communications between modules are provided by three 8 bit global video bus, and three local configurable pipeline 8 bit video bus. The VME bus is dedicated to system management. Tasks between DSPs are distributed as follows: two DSPs are used to edges detection, one for the right image and the other for the left one. The last processor computes the matching process and the 3D calculation. With $512{\times}512$ pixels images, this sensor generates dense 3D maps at a rate of about 1 Hz depending of the scene complexity. Results can surely be improved by using a special suited multiprocessors cards.

  • PDF

3D Reconstruction and Self-calibration based on Binocular Stereo Vision (스테레오 영상을 이용한 자기보정 및 3차원 형상 구현)

  • Hou, Rongrong;Jeong, Kyung-Seok
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.13 no.9
    • /
    • pp.3856-3863
    • /
    • 2012
  • A 3D reconstruction technique from stereo images that requires minimal intervention from the user has been developed. The reconstruction problem consists of three steps of estimating specific geometry groups. The first step is estimating the epipolar geometry that exists between the stereo image pairs which includes feature matching in both images. The second is estimating the affine geometry, a process to find a special plane in the projective space by means of vanishing points. The third step, which includes camera self-calibration, is obtaining a metric geometry from which a 3D model of the scene could be obtained. The major advantage of this method is that the stereo images do not need to be calibrated for reconstruction. The results of camera calibration and reconstruction have shown the possibility of obtaining a 3D model directly from features in the images.

Accuracy Investigation of DEM generated from Heterogeneous Stereo Satellite Images using Rational Polynomial Coefficients (RPC를 이용한 이종센서 위성영상으로부터의 수치고도모형 정확도 평가)

  • Song, Yeong-Sun
    • Journal of Korean Society for Geospatial Information Science
    • /
    • v.22 no.3
    • /
    • pp.121-128
    • /
    • 2014
  • This study investigated the accuracy of DEM generated by heterogeneous stereo satellite images based on RPC. Heterogeneous sensor images with different spatial resolution are SPOT-5 panchromatic and IKONOS images. For the accuracy evaluation of the DEM, we compared the DEMs generated from two kinds of sensors and that produced using homogeneous SPOT-5 and IKONOS stereo images. As results of the evaluation, accuracy of 3D positioning by heterogeneous images was substantially similar to that of homogeneous stereo images for exact conjugate points. But, in terms of quality of the DEM, DEM generated by heterogeneous sensor showed a lower accuracy about twice in RMSE and about 3 times in LE90 than that of homogeneous sensors. As a result, DEM can be generated by using heterogenous satellite imagery. But if we use a stereo image with different spatial resolution, the performance of image matching was very important factor for the production of high-quality DEM.

3-D Building Reconstruction from Standard IKONOS Stereo Products in Dense Urban Areas (IKONOS 컬러 입체영상을 이용한 대규모 도심지역의 3차원 건물복원)

  • Lee, Suk Kun;Park, Chung Hwan
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.26 no.3D
    • /
    • pp.535-540
    • /
    • 2006
  • This paper presented an effective strategy to extract the buildings and to reconstruct 3-D buildings using high-resolution multispectral stereo satellite images. Proposed scheme contained three major steps: building enhancement and segmentation using both BDT (Background Discriminant Transformation) and ISODATA algorithm, conjugate building identification using the object matching with Hausdorff distance and color indexing, and 3-D building reconstruction using photogrammetric techniques. IKONOS multispectral stereo images were used to evaluate the scheme. As a result, the BDT technique was verified as an effective tool for enhancing building areas since BDT suppressed the dominance of background to enhance the building as a non-background. In building recognition, color information itself was not enough to identify the conjugate building pairs since most buildings are composed of similar materials such as concrete. When both Hausdorff distance for edge information and color indexing for color information were combined, most segmented buildings in the stereo images were correctly identified. Finally, 3-D building models were successfully generated using the space intersection by the forward RFM (Rational Function Model).

Model-Based Plane Detection in Disparity Space Using Surface Partitioning (표면분할을 이용한 시차공간상에서의 모델 기반 평면검출)

  • Ha, Hong-joon;Lee, Chang-hun
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.4 no.10
    • /
    • pp.465-472
    • /
    • 2015
  • We propose a novel plane detection in disparity space and evaluate its performance. Our method simplifies and makes scenes in disparity space easily dealt with by approximating various surfaces as planes. Moreover, the approximated planes can be represented in the same size as in the real world, and can be employed for obstacle detection and camera pose estimation. Using a stereo matching technique, our method first creates a disparity image which consists of binocular disparity values at xy-coordinates in the image. Slants of disparity values are estimated by exploiting a line simplification algorithm which allows our method to reflect global changes against x or y axis. According to pairs of x and y slants, we label the disparity image. 4-connected disparities with the same label are grouped, on which least squared model estimates plane parameters. N plane models with the largest group of disparity values which satisfy their plane parameters are chosen. We quantitatively and qualitatively evaluate our plane detection. The result shows 97.9%와 86.6% of quality in our experiment respectively on cones and cylinders. Proposed method excellently extracts planes from Middlebury and KITTI dataset which are typically used for evaluation of stereo matching algorithms.

A Dynamic Programming Neural Network to find the Safety Distance of Industrial Field (산업 현장의 안전거리 계측을 위한 동적 계획 신경회로망)

  • Kim, Jong-Man;Kim, Won-Sub;Kim, Yeong-Min;Hwang, Jong-Sun;Park, Hyun-Chul
    • Proceedings of the Korean Institute of Electrical and Electronic Material Engineers Conference
    • /
    • 2001.09a
    • /
    • pp.23-27
    • /
    • 2001
  • Making the safety situation from the various work system is very important in the industrial fields. The proposed neural network technique is the real titre computation method based theory of inter-node diffusion for searching the safety distances from the sudden appearance-objests during the work driving. The main steps of the distance computation using the theory of stereo vision like the eyes of man is following steps. One is the processing for finding the corresponding points of stereo images and the other is the interpolation processing of full image data from nonlinear image data of obejects. All of them request much memory space and titre. Therefore the most reliable neural-network algorithm is drived for real time recognition of obejects, which is composed of a dynamic programming algorithm based on sequence matching techniques. And the real time reconstruction of nonlinear image information is processed through several simulations. I-D LIPN hardware has been composed, and the real time reconstruction is verified through the various experiments.

  • PDF

Relative Localization for Mobile Robot using 3D Reconstruction of Scale-Invariant Features (스케일불변 특징의 삼차원 재구성을 통한 이동 로봇의 상대위치추정)

  • Kil, Se-Kee;Lee, Jong-Shill;Ryu, Je-Goon;Lee, Eung-Hyuk;Hong, Seung-Hong;Shen, Dong-Fan
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.55 no.4
    • /
    • pp.173-180
    • /
    • 2006
  • A key component of autonomous navigation of intelligent home robot is localization and map building with recognized features from the environment. To validate this, accurate measurement of relative location between robot and features is essential. In this paper, we proposed relative localization algorithm based on 3D reconstruction of scale invariant features of two images which are captured from two parallel cameras. We captured two images from parallel cameras which are attached in front of robot and detect scale invariant features in each image using SIFT(scale invariant feature transform). Then, we performed matching for the two image's feature points and got the relative location using 3D reconstruction for the matched points. Stereo camera needs high precision of two camera's extrinsic and matching pixels in two camera image. Because we used two cameras which are different from stereo camera and scale invariant feature point and it's easy to setup the extrinsic parameter. Furthermore, 3D reconstruction does not need any other sensor. And the results can be simultaneously used by obstacle avoidance, map building and localization. We set 20cm the distance between two camera and capture the 3frames per second. The experimental results show :t6cm maximum error in the range of less than 2m and ${\pm}15cm$ maximum error in the range of between 2m and 4m.

3D feature point extraction technique using a mobile device (모바일 디바이스를 이용한 3차원 특징점 추출 기법)

  • Kim, Jin-Kyum;Seo, Young-Ho
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.256-257
    • /
    • 2022
  • In this paper, we introduce a method of extracting three-dimensional feature points through the movement of a single mobile device. Using a monocular camera, a 2D image is acquired according to the camera movement and a baseline is estimated. Perform stereo matching based on feature points. A feature point and a descriptor are acquired, and the feature point is matched. Using the matched feature points, the disparity is calculated and a depth value is generated. The 3D feature point is updated according to the camera movement. Finally, the feature point is reset at the time of scene change by using scene change detection. Through the above process, an average of 73.5% of additional storage space can be secured in the key point database. By applying the algorithm proposed to the depth ground truth value of the TUM Dataset and the RGB image, it was confirmed that the\re was an average distance difference of 26.88mm compared with the 3D feature point result.

  • PDF