• Title/Summary/Keyword: Camera calibration data

Search Result 167, Processing Time 0.024 seconds

Camera calibration parameters estimation using perspective variation ratio of grid type line widths (격자형 선폭들의 투영변화비를 이용한 카메라 교정 파라메터 추정)

  • Jeong, Jun-Ik;Choi, Seong-Gu;Rho, Do-Hwan
    • Proceedings of the KIEE Conference
    • /
    • 2004.11c
    • /
    • pp.30-32
    • /
    • 2004
  • With 3-D vision measuring, camera calibration is necessary to calculate parameters accurately. Camera calibration was developed widely in two categories. The first establishes reference points in space, and the second uses a grid type frame and statistical method. But, the former has difficulty to setup reference points and the latter has low accuracy. In this paper we present an algorithm for camera calibration using perspective ratio of the grid type frame with different line widths. It can easily estimate camera calibration parameters such as lens distortion, focal length, scale factor, pose, orientations, and distance. The advantage of this algorithm is that it can estimate the distance of the object. Also, the proposed camera calibration method is possible estimate distance in dynamic environment such as autonomous navigation. To validate proposed method, we set up the experiments with a frame on rotator at a distance of 1, 2, 3, 4[m] from camera and rotate the frame from -60 to 60 degrees. Both computer simulation and real data have been used to test the proposed method and very good results have been obtained. We have investigated the distance error affected by scale factor or different line widths and experimentally found an average scale factor that includes the least distance error with each image. The average scale factor tends to fluctuate with small variation and makes distance error decrease. Compared with classical methods that use stereo camera or two or three orthogonal planes, the proposed method is easy to use and flexible. It advances camera calibration one more step from static environments to real world such as autonomous land vehicle use.

  • PDF

The Structured Grid Calibration Based On Triangulation Method (삼각법을 기반으로 한 구조화된 격자 캘리브레이션)

  • KIM EUN-SEOK;JOO KI-SEE;WANG GI-NAM
    • Proceedings of the Korean Operations and Management Science Society Conference
    • /
    • 2002.05a
    • /
    • pp.248-252
    • /
    • 2002
  • Many sensors such as a structured grid pattern generator, a laser, and CCD camera to information have been used, but most of algorithms for a calibration are inefficient memory and experiment data are required. In this paper, the calibration algorithm of a structured grid pattern based on triang is introduced to calculate 3D information in the real world. The beams generated from str pattern generator established horizontally with the CCD camera are projected on the calibn CCD camera observes the intersection plane of a light and an object plane. The 3D infon calculated using observed and calibration data. This proposed method in this paper has advantages such as a memory saving and an experimental data since the 3D information are obtained simply triangulation method.

  • PDF

An Accurate Extrinsic Calibration of Laser Range Finder and Vision Camera Using 3D Edges of Multiple Planes (다중 평면의 3차원 모서리를 이용한 레이저 거리센서 및 카메라의 정밀 보정)

  • Choi, Sung-In;Park, Soon-Yong
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.4 no.4
    • /
    • pp.177-186
    • /
    • 2015
  • For data fusion of laser range finder (LRF) and vision camera, accurate calibration of external parameters which describe relative pose between two sensors is necessary. This paper proposes a new calibration method which can acquires more accurate external parameters between a LRF and a vision camera compared to other existing methods. The main motivation of the proposed method is that any corner data of a known 3D structure which is acquired by the LRF should be projected on a straight line in the camera image. To satisfy such constraint, we propose a 3D geometric model and a numerical solution to minimize the energy function of the model. In addition, we describe the implementation steps of the data acquisition of LRF and camera images which are necessary in accurate calibration results. In the experiment results, it is shown that the performance of the proposed method are better in terms of accuracy compared to other conventional methods.

The Camera Calibration Parameters Estimation using The Projection Variations of Line Widths (선폭들의 투영변화율을 이용한 카메라 교정 파라메터 추정)

  • Jeong, Jun-Ik;Moon, Sung-Young;Rho, Do-Hwan
    • Proceedings of the KIEE Conference
    • /
    • 2003.07d
    • /
    • pp.2372-2374
    • /
    • 2003
  • With 3-D vision measuring, camera calibration is necessary to calculate parameters accurately. Camera calibration was developed widely in two categories. The first establishes reference points in space, and the second uses a grid type frame and statistical method. But, the former has difficulty to setup reference points and the latter has low accuracy. In this paper we present an algorithm for camera calibration using perspective ratio of the grid type frame with different line widths. It can easily estimate camera calibration parameters such as focal length, scale factor, pose, orientations, and distance. But, radial lens distortion is not modeled. The advantage of this algorithm is that it can estimate the distance of the object. Also, the proposed camera calibration method is possible estimate distance in dynamic environment such as autonomous navigation. To validate proposed method, we set up the experiments with a frame on rotator at a distance of 1,2,3,4[m] from camera and rotate the frame from -60 to 60 degrees. Both computer simulation and real data have been used to test the proposed method and very good results have been obtained. We have investigated the distance error affected by scale factor or different line widths and experimentally found an average scale factor that includes the least distance error with each image. It advances camera calibration one more step from static environments to real world such as autonomous land vehicle use.

  • PDF

An Exact 3D Data Extraction Algorithm For Active Range Sensor using Laser Slit (레이저 슬릿을 사용하는 능동거리 센서의 정확한 3D 데이터 추출 알고리즘)

  • Cha, Y.Y.;Gweon, D.G.
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.12 no.8
    • /
    • pp.73-85
    • /
    • 1995
  • The sensor system to measure the distance precisely from the center of the sensor system to the obstacle is needed to recognize the surrounding environments, and the sensor system is to be calibrated thoroughly to get the range information exactly. This study covers the calibration of the active range sensor which consists of camera and laser slit emitting device, and provides the equations to get the 3D range data. This can be possible by obtaining the extrinsic parameters of laser slit emitting device through image processing the slits measured during the constant distance intervals and the intrinsic parameters from the calibration of camera. The 3D range data equation derived from the simple geometric assumptions is proved to be applicable to the general cases using the calibration parameters. Also the exact 3D range data were obtained to the object from the real experiment.

  • PDF

A 3D Foot Scanner Using Mirrors and Single Camera (거울 및 단일 카메라를 이용한 3차원 발 스캐너)

  • Chung, Seong-Youb;Park, Sang-Kun
    • Korean Journal of Computational Design and Engineering
    • /
    • v.16 no.1
    • /
    • pp.11-20
    • /
    • 2011
  • A structured beam laser is often used to scan object and make 3D model. Multiple cameras are inevitable to see occluded areas, which is the main reason of the high price of the scanner. In this paper, a low cost 3D foot scanner is developed using one camera and two mirrors. The camera and two mirrors are located below and above the foot, respectively. Occluded area, which is the top of the foot, is reflected by the mirrors. Then the camera measures 3D point data of the bottom and top of the foot at the same time. Then, the whole foot model is reconstructed after symmetrical transformation of the data reflected by mirrors. The reliability of the scan data depends on the accuracy of the parameters between the camera and the laser. A calibration method is also proposed and verified by experiments. The results of the experiments show that the worst errors of the system are 2 mm along x, y, and z directions.

3D Reconstruction using the Key-frame Selection from Reprojection Error (카메라 재투영 오차로부터 중요영상 선택을 이용한 3차원 재구성)

  • Seo, Yung-Ho;Kim, Sang-Hoon;Choi, Jong-Soo
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.45 no.1
    • /
    • pp.38-46
    • /
    • 2008
  • Key-frame selection algorithm is defined as the process of selecting a necessary images for 3D reconstruction from the uncalibrated images. Also, camera calibration of images is necessary for 3D reconstuction. In this paper, we propose a new method of Key-frame selection with the minimal error for camera calibration. Using the full-auto-calibration, we estimate camera parameters for all selected Key-frames. We remove the false matching using the fundamental matrix computed by algebraic deviation from the estimated camera parameters. Finally we obtain 3D reconstructed data. Our experimental results show that the proposed approach is required rather lower time costs than others, the error of reconstructed data is the smallest. The elapsed time for estimating the fundamental matrix is very fast and the error of estimated fundamental matrix is similar to others.

Development of Color 3D Scanner Using Laser Structured-light Imaging Method

  • Ko, Youngjun;Yi, Sooyeong
    • Current Optics and Photonics
    • /
    • v.2 no.6
    • /
    • pp.554-562
    • /
    • 2018
  • This study presents a color 3D scanner based on the laser structured-light imaging method that can simultaneously acquire 3D shape data and color of a target object using a single camera. The 3D data acquisition of the scanner is based on the structured-light imaging method, and the color data is obtained from a natural color image. Because both the laser image and the color image are acquired by the same camera, it is efficient to obtain the 3D data and the color data of a pixel by avoiding the complicated correspondence algorithm. In addition to the 3D data, the color data is helpful for enhancing the realism of an object model. The proposed scanner consists of two line lasers, a color camera, and a rotation table. The line lasers are deployed at either side of the camera to eliminate shadow areas of a target object. This study addresses the calibration methods for the parameters of the camera, the plane equations covered by the line lasers, and the center of the rotation table. Experimental results demonstrate the performance in terms of accurate color and 3D data acquisition in this study.

Motion Analysis of a Moving Object using one Camera and Tracking Method (단일 카메라와 Tracking 기법을 이용한 이동 물체의 모션 분석)

  • Shin, Myong-Jun;Son, Young-Ik;Kim, Kab-Il
    • Proceedings of the KIEE Conference
    • /
    • 2005.07d
    • /
    • pp.2821-2823
    • /
    • 2005
  • When we deal with the image data through camera lens, much works are necessary for removing image distortions and obtaining accurate informations from the raw data. However, the calibration process is very complicated and requires many trials and errors. In this paper, 3 new approach to image processing is presented by developing a H/W vision system with a tracking camera. Using motor control with encoders the proposed tracking method tells us exact displacements of a moving object. Therefore this method does not require any calibration process for pin cusion. Owing to the mobility one camera covers wide ranges and, by lowering its height, the camera also obtains high resolution of the image. We first introduce the structure of the motion analysis system. Then the construced vision system is investigated by some experiments.

  • PDF

Modeling and Calibration of a 3D Robot Laser Scanning System (3차원 로봇 레이저 스캐닝 시스템의 모델링과 캘리브레이션)

  • Lee Jong-Kwang;Yoon Ji Sup;Kang E-Sok
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.11 no.1
    • /
    • pp.34-40
    • /
    • 2005
  • In this paper, we describe the modeling for the 3D robot laser scanning system consisting of a laser stripe projector, camera, and 5-DOF robot and propose its calibration method. Nonlinear radial distortion in the camera model is considered for improving the calibration accuracy. The 3D range data is calculated using the optical triangulation principle which uses the geometrical relationship between the camera and the laser stripe plane. For optimal estimation of the system model parameters, real-coded genetic algorithm is applied in the calibration process. Experimental results show that the constructed system is able to measure the 3D position within about 1mm error. The proposed scheme could be applied to the kinematically dissimilar robot system without losing the generality and has a potential for recognition for the unknown environment.