• Title/Summary/Keyword: camera motion parameter

Search Result 45, Processing Time 0.027 seconds

A Defocus Technique based Depth from Lens Translation using Sequential SVD Factorization

  • Kim, Jong-Il;Ahn, Hyun-Sik;Jeong, Gu-Min;Kim, Do-Hyun
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.383-388
    • /
    • 2005
  • Depth recovery in robot vision is an essential problem to infer the three dimensional geometry of scenes from a sequence of the two dimensional images. In the past, many studies have been proposed for the depth estimation such as stereopsis, motion parallax and blurring phenomena. Among cues for depth estimation, depth from lens translation is based on shape from motion by using feature points. This approach is derived from the correspondence of feature points detected in images and performs the depth estimation that uses information on the motion of feature points. The approaches using motion vectors suffer from the occlusion or missing part problem, and the image blur is ignored in the feature point detection. This paper presents a novel approach to the defocus technique based depth from lens translation using sequential SVD factorization. Solving such the problems requires modeling of mutual relationship between the light and optics until reaching the image plane. For this mutuality, we first discuss the optical properties of a camera system, because the image blur varies according to camera parameter settings. The camera system accounts for the camera model integrating a thin lens based camera model to explain the light and optical properties and a perspective projection camera model to explain the depth from lens translation. Then, depth from lens translation is proposed to use the feature points detected in edges of the image blur. The feature points contain the depth information derived from an amount of blur of width. The shape and motion can be estimated from the motion of feature points. This method uses the sequential SVD factorization to represent the orthogonal matrices that are singular value decomposition. Some experiments have been performed with a sequence of real and synthetic images comparing the presented method with the depth from lens translation. Experimental results have demonstrated the validity and shown the applicability of the proposed method to the depth estimation.

  • PDF

Dynamic tracking control of robot manipulators using vision system (비전 시스템을 이용한 로봇 머니퓰레이터의 동력학 추적 제어)

  • 한웅기;국태용
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1997.10a
    • /
    • pp.1816-1819
    • /
    • 1997
  • Using the vision system, robotic tasks in unstructured environments can be accompished, which reduces greatly the cost and steup time for the robotic system to fit to he well-defined and structured working environments. This paper proposes a dynamic control scheme for robot manipulator with eye-in-hand camera configuration. To perfom the tasks defined in the image plane, the camera motion Jacobian (image Jacobian) matrix is used to transform the camera motion to the objection position change. In addition, the dynamic learning controller is designed to improve the tracking performance of robotic system. the proposed control scheme is implemented for tasks of tracking moving objects and shown to outperform the conventional visual servo system in convergence and robustness to parameter uncertainty, disturbances, low sampling rate, etc.

  • PDF

Camera Motion Estimation using Geometrically Symmetric Points in Subsequent Video Frames (인접 영상 프레임에서 기하학적 대칭점을 이용한 카메라 움직임 추정)

  • Jeon, Dae-Seong;Mun, Seong-Heon;Park, Jun-Ho;Yun, Yeong-U
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.39 no.2
    • /
    • pp.35-44
    • /
    • 2002
  • The translation and the rotation of camera occur global motion which affects all over the frame in video sequence. With the video sequences containing global motion, it is practically impossible to extract exact video objects and to calculate genuine object motions. Therefore, high compression ratio cannot be achieved due to the large motion vectors. This problem can be solved when the global motion compensated frames are used. The existing camera motion estimation methods for global motion compensation have a large amount of computations in common. In this paper, we propose a simple global motion estimation algorithm that consists of linear equations without any repetition. The algorithm uses information .of symmetric points in the frame of the video sequence. The discriminant conditions to distinguish regions belonging to distant view from foreground in the frame are presented. Only for the distant view satisfying the discriminant conditions, the linear equations for the panning, tilting, and zooming parameters are applied. From the experimental results using the MPEG test sequences, we can confirm that the proposed algorithm estimates correct global motion parameters. Moreover the real-time capability of the proposed technique can be applicable to many MPEG-4 and MPEG-7 related areas.

Camera Extrinsic Parameter Estimation using 2D Homography and Nonlinear Minimizing Method based on Geometric Invariance Vector (기하학적 불변벡터 기탄 2D 호모그래피와 비선형 최소화기법을 이용한 카메라 외부인수 측정)

  • Cha, Jeong-Hee
    • Journal of Internet Computing and Services
    • /
    • v.6 no.6
    • /
    • pp.187-197
    • /
    • 2005
  • In this paper, we propose a method to estimate camera motion parameter based on invariant point features, Typically, feature information of image has drawbacks, it is variable to camera viewpoint, and therefore information quantity increases after time, The LM(Levenberg-Marquardt) method using nonlinear minimum square evaluation for camera extrinsic parameter estimation also has a weak point, which has different iteration number for approaching the minimal point according to the initial values and convergence time increases if the process run into a local minimum, In order to complement these shortfalls, we, first proposed constructing feature models using invariant vector of geometry, Secondly, we proposed a two-stage calculation method to improve accuracy and convergence by using 2D homography and LM method, In the experiment, we compared and analyzed the proposed method with existing method to demonstrate the superiority of the proposed algorithms.

  • PDF

Moving Target Tracking Algorithm based on the Confidence Measure of Motion Vectors (움직임 벡터의 신뢰도에 기반한 이동 목표물 추적 기법)

  • Lee, Jin-Seong;Lee, Gwang-Yeon;Kim, Seong-Dae
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.38 no.2
    • /
    • pp.160-168
    • /
    • 2001
  • Change detection using difference picture has been used to detect the location of moving targets and to track them. This method needs the assumption of static camera, and the global motion compensation is required in case of a moving camera. This paper suggests a method for finding a minimum bounding rectangles(MBR) of moving targets in the image sequences using moving region detection, especially with a moving camera. If the global motion parameter is inaccurately estimated, the estimated locations of targets will be accurate either To alleviate this problem, we introduce the concept of the confidence measure and achieve more accurate estimation of global motion. Experimental results show that the proposed method successfully removes background region and extracts MBRs of the targets. Even with a moving camera, the new global motion estimation algorithm performs more precise]y and it reduces the background compensation errors of change detection.

  • PDF

LiDAR Data Interpolation Algorithm for 3D-2D Motion Estimation (3D-2D 모션 추정을 위한 LiDAR 정보 보간 알고리즘)

  • Jeon, Hyun Ho;Ko, Yun Ho
    • Journal of Korea Multimedia Society
    • /
    • v.20 no.12
    • /
    • pp.1865-1873
    • /
    • 2017
  • The feature-based visual SLAM requires 3D positions for the extracted feature points to perform 3D-2D motion estimation. LiDAR can provide reliable and accurate 3D position information with low computational burden, while stereo camera has the problem of the impossibility of stereo matching in simple texture image region, the inaccuracy in depth value due to error contained in intrinsic and extrinsic camera parameter, and the limited number of depth value restricted by permissible stereo disparity. However, the sparsity of LiDAR data may increase the inaccuracy of motion estimation and can even lead to the result of motion estimation failure. Therefore, in this paper, we propose three interpolation methods which can be applied to interpolate sparse LiDAR data. Simulation results obtained by applying these three methods to a visual odometry algorithm demonstrates that the selective bilinear interpolation shows better performance in the view point of computation speed and accuracy.

Multi-camera Calibration Method for Optical Motion Capture System (광학식 모션캡처를 위한 다중 카메라 보정 방법)

  • Shin, Ki-Young;Mun, Joung-H.
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.6
    • /
    • pp.41-49
    • /
    • 2009
  • In this paper, the multi-camera calibration algorithm for optical motion capture system is proposed. This algorithm performs 1st camera calibration using DLT(Direct linear transformation} method and 3-axis calibration frame with 7 optical markers. And 2nd calibration is performed by waving with a wand of known length(so called wand dance} throughout desired calibration volume. In the 1st camera calibration, it is obtained not only camera parameter but also radial lens distortion parameters. These parameters are used initial solution for optimization in the 2nd camera calibration. In the 2nd camera calibration, the optimization is performed. The objective function is to minimize the difference of distance between real markers and reconstructed markers. For verification of the proposed algorithm, re-projection errors are calculated and the distance among markers in the 3-axis frame and in the wand calculated. And then it compares the proposed algorithm with commercial motion capture system. In the 3D reconstruction error of 3-axis frame, average error presents 1.7042mm(commercial system) and 0.8765mm(proposed algorithm). Average error reduces to 51.4 percent in commercial system. In the distance between markers in the wand, the average error shows 1.8897mm in the commercial system and 2.0183mm in the proposed algorithm.

A Study on The Identification of Blur Parameters from a Motion Blurred Image (모션 블러된 이미지로부터 블러 파라미터를 추출하는 기법에 대한 연구)

  • Yang, Hong-Taek;Hwang, Joo-Yeon;Paik, Doo-Won
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02a
    • /
    • pp.693-696
    • /
    • 2008
  • Motion blurs are caused by relative motion between the camera and the scene. The blurred image needs to be restored because undesired blur effect degrades the quality of the image. In this paper, we propose a new method for the identification of blur parameters. Experiment shows that the proposed method identifies blur extent regardless of the size of the blur and the object in the original image.

  • PDF

Development of TPF Generation SIW for KOMPSAT-2 X-Band Antenna Motion Control

  • Kang C. H.;Park D. J.;Seo S. B.;Koo I. H.;Ahn S. I.;Kim E. K.
    • Proceedings of the KSRS Conference
    • /
    • 2005.10a
    • /
    • pp.485-488
    • /
    • 2005
  • The 2nd KOrea Multi-Purpose Satellite (KOMPSAT -2) has been developed by Korea Aerospace Research Institute (KARI) since 2000. Multi Spectral Camera (MSC) is the payload for KOMPSAT -2, which will provide the observation imagery around Korean peninsula with high resolution. KOMPSAT-2 has adopted X-band Tracking System (XTS) for transmitting earth observation data to ground station. For this, data which describes and controls the pre-defined motion of each on-board X-Band antenna in XTS, must be transmitted to the spacecraft as S-Band command and it is called as Tracking Parameter Files (TPF). In this paper, the result of the development of TPF Generation S/W for KOMPSAT-2 X-Band Antenna Motion Control.

  • PDF

Spatial Multilevel Optical Flow Architecture-based Dynamic Motion Estimation in Vehicular Traffic Scenarios

  • Fuentes, Alvaro;Yoon, Sook;Park, Dong Sun
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.12
    • /
    • pp.5978-5999
    • /
    • 2018
  • Pedestrian detection is a challenging area in the intelligent vehicles domain. During the last years, many works have been proposed to efficiently detect motion in images. However, the problem becomes more complex when it comes to detecting moving areas while the vehicle is also moving. This paper presents a variational optical flow-based method for motion estimation in vehicular traffic scenarios. We introduce a framework for detecting motion areas with small and large displacements by computing optical flow using a multilevel architecture. The flow field is estimated at the shortest level and then successively computed until the largest level. We include a filtering parameter and a warping process using bicubic interpolation to combine the intermediate flow fields computed at each level during optimization to gain better performance. Furthermore, we find that by including a penalization function, our system is able to effectively reduce the presence of outliers and deal with all expected circumstances in real scenes. Experimental results are performed on various image sequences from Daimler Pedestrian Dataset that includes urban traffic scenarios. Our evaluation demonstrates that despite the complexity of the evaluated scenes, the motion areas with both moving and static camera can be effectively identified.