• Title/Summary/Keyword: Distortion Invariant

Search Result 49, Processing Time 0.027 seconds

Multi-view Image Generation from Stereoscopic Image Features and the Occlusion Region Extraction (가려짐 영역 검출 및 스테레오 영상 내의 특징들을 이용한 다시점 영상 생성)

  • Lee, Wang-Ro;Ko, Min-Soo;Um, Gi-Mun;Cheong, Won-Sik;Hur, Nam-Ho;Yoo, Ji-Sang
    • Journal of Broadcast Engineering
    • /
    • v.17 no.5
    • /
    • pp.838-850
    • /
    • 2012
  • In this paper, we propose a novel algorithm that generates multi-view images by using various image features obtained from the given stereoscopic images. In the proposed algorithm, we first create an intensity gradient saliency map from the given stereo images. And then we calculate a block-based optical flow that represents the relative movement(disparity) of each block with certain size between left and right images. And we also obtain the disparities of feature points that are extracted by SIFT(scale-invariant We then create a disparity saliency map by combining these extracted disparity features. Disparity saliency map is refined through the occlusion detection and removal of false disparities. Thirdly, we extract straight line segments in order to minimize the distortion of straight lines during the image warping. Finally, we generate multi-view images by grid mesh-based image warping algorithm. Extracted image features are used as constraints during grid mesh-based image warping. The experimental results show that the proposed algorithm performs better than the conventional DIBR algorithm in terms of visual quality.

Mixed Mobile Education System using SIFT Algorithm (SIFT 알고리즘을 이용한 혼합형 모바일 교육 시스템)

  • Hong, Kwang-Jin;Jung, Kee-Chul;Han, Eun-Jung;Yang, Jong-Yeol
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.13 no.2
    • /
    • pp.69-79
    • /
    • 2008
  • Due to popularization of the wireless Internet and mobile devices the infrastructure of the ubiquitous environment, where users can get information whatever they want anytime and anywhere, is created. Therefore, a variety of fields including the education studies methods for efficiency of information transmission using on-line and off-line contents. In this paper, we propose the Mixed Mobile Education system(MME) that improves educational efficiency using on-line and off-line contents on mobile devices. Because it is hard to input new data and cannot use similar off-line contents in systems used additional tags, the proposed system does not use additional tags but recognizes of-line contents as we extract feature points in the input image using the mobile camera. We use the Scale Invariant Feature Transform(SIFT) algorithm to extract feature points which are not affected by noise, color distortion, size and rotation in the input image captured by the low resolution camera. And we use the client-server architecture for solving the limited storage size of the mobile devices and for easily registration and modification of data. Experimental results show that compared with previous work, the proposed system has some advantages and disadvantages and that the proposed system has good efficiency on various environments.

  • PDF

Comparison of Multi-angle TerraSAR-X Staring Mode Image Registration Method through Coarse to Fine Step (Coarse to Fine 단계를 통한 TerraSAR-X Staring Mode 다중 관측각 영상 정합기법 비교 분석)

  • Lee, Dongjun;Kim, Sang-Wan
    • Korean Journal of Remote Sensing
    • /
    • v.37 no.3
    • /
    • pp.475-491
    • /
    • 2021
  • With the recent increase in available high-resolution (< ~1 m) satellite SAR images, the demand for precise registration of SAR images is increasing in various fields including change detection. The registration between high-resolution SAR images acquired in different look angle is difficult due to speckle noise and geometric distortion caused by the characteristics of SAR images. In this study, registration is performed in two stages, coarse and fine, using the x-band SAR data imaged at staring spotlight mode of TerraSAR-X. For the coarse registration, a method combining the adaptive sampling method and SAR-SIFT (Scale Invariant Feature Transform) is applied, and three rigid methods (NCC: Normalized Cross Correlation, Phase Congruency-NCC, MI: Mutual Information) and one non-rigid (Gefolki: Geoscience extended Flow Optical Flow Lucas-Kanade Iterative), for the fine registration stage, was performed for performance comparison. The results were compared by using RMSE (Root Mean Square Error) and FSIM (Feature Similarity) index, and all rigid models showed poor results in all image combinations. It is confirmed that the rigid models have a large registration error in the rugged terrain area. As a result of applying the Gefolki algorithm, it was confirmed that the RMSE of Gefolki showed the best result as a 1~3 pixels, and the FSIM index also obtained a higher value than 0.02~0.03 compared to other rigid methods. It was confirmed that the mis-registration due to terrain effect could be sufficiently reduced by the Gefolki algorithm.

Comparison of Feature Point Extraction Algorithms Using Unmanned Aerial Vehicle RGB Reference Orthophoto (무인항공기 RGB 기준 정사영상을 이용한 특징점 추출 알고리즘 비교)

  • Lee, Kirim;Seong, Jihoon;Jung, Sejung;Shin, Hyeongil;Kim, Dohoon;Lee, Wonhee
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.44 no.2
    • /
    • pp.263-270
    • /
    • 2024
  • As unmanned aerial vehicles(UAVs) and sensors have been developed in a variety of ways, it has become possible to update information on the ground faster than existing aerial photography or remote sensing. However, acquisition and input of ground control points(GCPs) UAV photogrammetry takes a lot of time, and geometric distortion occurs if measurement and input of GCPs are incorrect. In this study, RGB-based orthophotos were generated to reduce GCPs measurment and input time, and comparison and evaluation were performed by applying feature point algorithms to target orthophotos from various sensors. Four feature point extraction algorithms were applied to the two study sites, and as a result, speeded up robust features(SURF) was the best in terms of the ratio of matching pairs to feature points. When compared overall, the accelerated-KAZE(AKAZE) method extracted the most feature points and matching pairs, and the binary robust invariant scalable keypoints(BRISK) method extracted the fewest feature points and matching pairs. Through these results, it was confirmed that the AKAZE method is superior when performing geometric correction of the objective orthophoto for each sensor.

A panorama image generation method using FAST algorithm (FAST를 이용한 파노라마 영상 생성 방법)

  • Kim, Jong-ho;Ko, Jin-woong;Yoo, Jisang
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.20 no.3
    • /
    • pp.630-638
    • /
    • 2016
  • In this paper, a feature based panorama image generation algorithm using FAST(Features from Accelerated Segment Test) method that is faster than SIFT(Scale Invariant Feature Transform) and SURF(Speeded Up Robust Features) is proposed. Cylindrical projection is performed to generate natural panorama images with numerous images as input. The occurred error can be minimized by applying RANSAC(Random Sample Consensus) for the matching process. When we synthesize numerous images acquired from different camera angles, we use blending techniques to compensate the distortions by the heterogeneity of border line. In that way, we could get more natural synthesized panorama image. The proposed algorithm can generate natural panorama images regardless the order of input images and tilted images. In addition, the image matching can be faster than the conventional method. As a result of the experiments, distortion was corrected and natural panorama image was generated.

Face Recognition under Varying Pose using Local Area obtained by Side-view Pose Normalization (측면 포즈정규화를 통한 부분 영역을 이용한 포즈 변화에 강인한 얼굴 인식)

  • Ahn, Byeong-Doo;Ko, Han-Seok
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.42 no.4 s.304
    • /
    • pp.59-68
    • /
    • 2005
  • This paper proposes a face recognition under varying poses using local area obtained by side-view pose normalization. General normalization methods for face recognition under varying pose have a problem with the information about invisible area of face. Generally this problem is solved by compensation, but there are many cases where the image is distorted or features lost due to compensation .To solve this problem we normalize the face pose in side-view to reduce distortion that happens mainly in areas that have large depth variation. We only use undistorted area, removing the area that has been distorted by normalization. We consider two cases of yaw pose variation and pitch pose variation, and by experiments, we confirm the improvement of recognition performance.

GAN-based Image-to-image Translation using Multi-scale Images (다중 스케일 영상을 이용한 GAN 기반 영상 간 변환 기법)

  • Chung, Soyoung;Chung, Min Gyo
    • The Journal of the Convergence on Culture Technology
    • /
    • v.6 no.4
    • /
    • pp.767-776
    • /
    • 2020
  • GcGAN is a deep learning model to translate styles between images under geometric consistency constraint. However, GcGAN has a disadvantage that it does not properly maintain detailed content of an image, since it preserves the content of the image through limited geometric transformation such as rotation or flip. Therefore, in this study, we propose a new image-to-image translation method, MSGcGAN(Multi-Scale GcGAN), which improves this disadvantage. MSGcGAN, an extended model of GcGAN, performs style translation between images in a direction to reduce semantic distortion of images and maintain detailed content by learning multi-scale images simultaneously and extracting scale-invariant features. The experimental results showed that MSGcGAN was better than GcGAN in both quantitative and qualitative aspects, and it translated the style more naturally while maintaining the overall content of the image.

A partially occluded object recognition technique using a probabilistic analysis in the feature space (특징 공간상에서 의 확률적 해석에 기반한 부분 인식 기법에 관한 연구)

  • 박보건;이경무;이상욱;이진학
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.26 no.11A
    • /
    • pp.1946-1956
    • /
    • 2001
  • In this paper, we propose a novel 2-D partial matching algorithm based on model-based stochastic analysis of feature correspondences in a relation vector space, which is quite robust to shape variations as well as invariant to geometric transformations. We represent an object using the ARG (Attributed Relational Graph) model with features of a set of relation vectors. In addition, we statistically model the partial occlusion or noise as the distortion of the relation vector distribution in the relation vector space. Our partial matching algorithm consists of two-phases. First, a finite number of candidate sets areselected by using logical constraint embedding local and structural consistency Second, the feature loss detection is done iteratively by error detection and voting scheme thorough the error analysis of relation vector space. Experimental results on real images demonstrate that the proposed algorithm is quite robust to noise and localize target objects correctly even inseverely noisy and occluded scenes.

  • PDF

Vision-based Mobile Robot Localization and Mapping using fisheye Lens (어안렌즈를 이용한 비전 기반의 이동 로봇 위치 추정 및 매핑)

  • Lee Jong-Shill;Min Hong-Ki;Hong Seung-Hong
    • Journal of the Institute of Convergence Signal Processing
    • /
    • v.5 no.4
    • /
    • pp.256-262
    • /
    • 2004
  • A key component of an autonomous mobile robot is to localize itself and build a map of the environment simultaneously. In this paper, we propose a vision-based localization and mapping algorithm of mobile robot using fisheye lens. To acquire high-level features with scale invariance, a camera with fisheye lens facing toward to ceiling is attached to the robot. These features are used in mP building and localization. As a preprocessing, input image from fisheye lens is calibrated to remove radial distortion and then labeling and convex hull techniques are used to segment ceiling and wall region for the calibrated image. At the initial map building process, features we calculated for each segmented region and stored in map database. Features are continuously calculated for sequential input images and matched to the map. n some features are not matched, those features are added to the map. This map matching and updating process is continued until map building process is finished, Localization is used in map building process and searching the location of the robot on the map. The calculated features at the position of the robot are matched to the existing map to estimate the real position of the robot, and map building database is updated at the same time. By the proposed method, the elapsed time for map building is within 2 minutes for 50㎡ region, the positioning accuracy is ±13cm and the error about the positioning angle of the robot is ±3 degree for localization.

  • PDF