• Title/Summary/Keyword: 어안렌즈 교정

Search Result 4, Processing Time 0.02 seconds

Calibration of Fisheye Lens Images Using a Spiral Pattern and Compensation for Geometric Distortion (나선형 패턴을 사용한 어안렌즈 영상 교정 및 기하학적 왜곡 보정)

  • Kim, Seon-Yung;Yoon, In-Hye;Kim, Dong-Gyun;Paik, Joon-Ki
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.49 no.4
    • /
    • pp.16-22
    • /
    • 2012
  • In this paper, we present spiral pattern which suits for optical simulator to calibrate fisheye lens and compensate geometric distortion. Using spiral pattern, we present calibration without mathematical modeling in advance. Proposed spiral pattern used to input image of optical simulator. Using fisheye lens image, we calibrate a fisheye lens by matching geometrically moved dots to corresponding original dots which leads not to need mathematical modeling. Proposed algorithm calibrates using dot matching which matches spiral pattern image dot to distorted image dot. And this algorithm does not need modeling in advance so it is effective. Proposed algorithm is enabled at processing of pattern recognition which has to get the exact information using fisheye lens for digital zooming. And this makes possible at compensation of geometric distortion and calibration of fisheye lens image applying in various image processing.

Verification Method of Omnidirectional Camera Model by Projected Contours (사영된 컨투어를 이용한 전방향 카메라 모델의 검증 방법)

  • Hwang, Yong-Ho;Lee, Jae-Man;Hong, Hyun-Ki
    • 한국HCI학회:학술대회논문집
    • /
    • 2007.02a
    • /
    • pp.994-999
    • /
    • 2007
  • 전방향(omnidirectional) 카메라 시스템은 보다 적은 수의 영상으로부터 주변 장면(scene)에 대한 많은 정보를 취득할 수 있는 장점이 있기 때문에 전방향 영상을 이용한 자동교정(self-calibration)과 3차원 재구성 등의 연구가 활발히 진행되고 있다. 본 논문에서는 기존에 제안된 교정 방법들을 이용하여 추정된 사영모델(projection model)의 정확성을 검증하기 위한 새로운 방법이 제안된다. 실 세계에서 다양하게 존재하는 직선 성분들은 전방향 영상에 컨투어(contour)의 형태로 사영되며, 사영모델과 컨투어의 양 끝점 좌표 값을 이용하여 그 궤적을 추정할 수 있다. 추정된 컨투어의 궤적과 영상에 존재하는 컨투어와의 거리 오차(distance error)로부터 전방향 카메라의 사영모델의 정확성을 검증할 수 있다. 제안된 방법의 성능을 평가하기 위해서 구 맵핑(spherical mapping)된 합성(synthetic) 영상과 어안렌즈(fisheye lens)로 취득한 실제 영상에 대해 제안된 알고리즘을 적용하여 사영모델의 정확성을 판단하였다.

  • PDF

3D Analysis of Scene and Light Environment Reconstruction for Image Synthesis (영상합성을 위한 3D 공간 해석 및 조명환경의 재구성)

  • Hwang, Yong-Ho;Hong, Hyun-Ki
    • Journal of Korea Game Society
    • /
    • v.6 no.2
    • /
    • pp.45-50
    • /
    • 2006
  • In order to generate a photo-realistic synthesized image, we should reconstruct light environment by 3D analysis of scene. This paper presents a novel method for identifying the positions and characteristics of the lights-the global and local lights-in the real image, which are used to illuminate the synthetic objects. First, we generate High Dynamic Range(HDR) radiance map from omni-directional images taken by a digital camera with a fisheye lens. Then, the positions of the camera and light sources in the scene are identified automatically from the correspondences between images without a priori camera calibration. Types of the light sources are classified according to whether they illuminate the whole scene, and then we reconstruct 3D illumination environment. Experimental results showed that the proposed method with distributed ray tracing makes it possible to achieve photo-realistic image synthesis. It is expected that animators and lighting experts for the film and animation industry would benefit highly from it.

  • PDF

Omnidirectional Camera Motion Estimation Using Projected Contours (사영 컨투어를 이용한 전방향 카메라의 움직임 추정 방법)

  • Hwang, Yong-Ho;Lee, Jae-Man;Hong, Hyun-Ki
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.44 no.5
    • /
    • pp.35-44
    • /
    • 2007
  • Since the omnidirectional camera system with a very large field of view could take many information about environment scene from few images, various researches for calibration and 3D reconstruction using omnidirectional image have been presented actively. Most of line segments of man-made objects we projected to the contours by using the omnidirectional camera model. Therefore, the corresponding contours among images sequences would be useful for computing the camera transformations including rotation and translation. This paper presents a novel two step minimization method to estimate the extrinsic parameters of the camera from the corresponding contours. In the first step, coarse camera parameters are estimated by minimizing an angular error function between epipolar planes and back-projected vectors from each corresponding point. Then we can compute the final parameters minimizing a distance error of the projected contours and the actual contours. Simulation results on the synthetic and real images demonstrated that our algorithm can achieve precise contour matching and camera motion estimation.