• Title/Summary/Keyword: SIFT

Search Result 345, Processing Time 0.029 seconds

Region based Scene Segmentation method for Topography Analysis (지형 분석을 위한 영역 기반 장면 분할 기법)

  • Jeon, Taegyun;Jeon, Moongu
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2012.11a
    • /
    • pp.503-506
    • /
    • 2012
  • 본 연구에서는 일반 야외 영상 및 항공 시뮬레이션 영상에 대한 지형 분석을 위해 영역 기반 장면 분할 기법을 제시한다. 영역의 분류를 위해 MeanShift 기법을 기반으로 한 표현과 Texton, SIFT, 위치정보를 특징으로 하는 기법을 제안하고 실험을 통해 주요 대상 영역이 분할되는 결과를 보인다. Sowerby 데이터 셋과 Google Earth 데이터로부터 자체적으로 제작한 데이터 셋에 대해 실험하였으며 수풀지형, 초목지형, 도로 등에 대해 분류하였다.

Implementation of the Panoramic System Using Feature-Based Image Stitching (특징점 기반 이미지 스티칭을 이용한 파노라마 시스템 구현)

  • Choi, Jaehak;Lee, Yonghwan;Kim, Youngseop
    • Journal of the Semiconductor & Display Technology
    • /
    • v.16 no.2
    • /
    • pp.61-65
    • /
    • 2017
  • Recently, the interest and research on 360 camera and 360 image production are expanding. In this paper, we describe the feature extraction algorithm, alignment and image blending that make up the feature-based stitching system. And it deals with the theory of representative algorithm at each stage. In addition, the feature-based stitching system was implemented using OPENCV library. As a result of the implementation, the brightness of the two images is different, and it feels a sense of heterogeneity in the resulting image. We will study the proper preprocessing to adjust the brightness value to improve the accuracy and seamlessness of the feature-based stitching system.

  • PDF

Metadata Processing Technique for Similar Image Search of Mobile Platform

  • Seo, Jung-Hee
    • Journal of information and communication convergence engineering
    • /
    • v.19 no.1
    • /
    • pp.36-41
    • /
    • 2021
  • Text-based image retrieval is not only cumbersome as it requires the manual input of keywords by the user, but is also limited in the semantic approach of keywords. However, content-based image retrieval enables visual processing by a computer to solve the problems of text retrieval more fundamentally. Vision applications such as extraction and mapping of image characteristics, require the processing of a large amount of data in a mobile environment, rendering efficient power consumption difficult. Hence, an effective image retrieval method on mobile platforms is proposed herein. To provide the visual meaning of keywords to be inserted into images, the efficiency of image retrieval is improved by extracting keywords of exchangeable image file format metadata from images retrieved through a content-based similar image retrieval method and then adding automatic keywords to images captured on mobile devices. Additionally, users can manually add or modify keywords to the image metadata.

Remote Sensing Image Registration using Structure Extraction and Keypoint Filtering (구조물 검출 네트워크 및 특징점 필터링을 이용한 원격 탐사 영상 정합)

  • Sung, Jun-Young;Lee, Woo-Ju;Oh, Seoung-Jun
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2020.07a
    • /
    • pp.300-304
    • /
    • 2020
  • 본 논문에서는 원격 탐사 영상 정합에서 정확도는 유지하면서 특징점 매칭 (Matching) 복잡도를 줄이기 위해 입력 영상을 전처리하는 구조물 검출 네트워크를 이용한 원격 탐사 영상 정합 방법을 제안한다. 영상 정합의 기존 방법은 입력 영상에서 특징점을 추출하고 설명자 (Descriptor)를 생성한다. 본 논문에서 제안하는 방법은 입력 영상에서 특징점 매칭에 영향을 미치는 구조물만 추출하여 새로운 영상을 만들어 특징점을 추출한다. 추출된 특징점은 필터링 (Filtering)을 거쳐 원본 영상에 매핑 (Mapping)되어 설명자를 생성하여 특징점 매칭 속도를 향상시킨다. 또한 구조물 검출 네트워크에서 학습 영상과 시험 영상의 특성의 차이로 생기는 성능 저하 문제를 개선하기 위해 히스토그램 매핑 기법을 이용한다. 아리랑 3 호가 획득한 원격 탐사 영상에 대한 실험을 통해 제안하는 방법은 정확도를 유지하면서 계산 시간을 SURF 보다 87.5%, SIFT 보다 92.6% 감소시킬 수 있다.

  • PDF

BoF based Action Recognition using Spatio-Temporal 2D Descriptor (시공간 2D 특징 설명자를 사용한 BOF 방식의 동작인식)

  • KIM, JinOk
    • Journal of Internet Computing and Services
    • /
    • v.16 no.3
    • /
    • pp.21-32
    • /
    • 2015
  • Since spatio-temporal local features for video representation have become an important issue of modeless bottom-up approaches in action recognition, various methods for feature extraction and description have been proposed in many papers. In particular, BoF(bag of features) has been promised coherent recognition results. The most important part for BoF is how to represent dynamic information of actions in videos. Most of existing BoF methods consider the video as a spatio-temporal volume and describe neighboring 3D interest points as complex volumetric patches. To simplify these complex 3D methods, this paper proposes a novel method that builds BoF representation as a way to learn 2D interest points directly from video data. The basic idea of proposed method is to gather feature points not only from 2D xy spatial planes of traditional frames, but from the 2D time axis called spatio-temporal frame as well. Such spatial-temporal features are able to capture dynamic information from the action videos and are well-suited to recognize human actions without need of 3D extensions for the feature descriptors. The spatio-temporal BoF approach using SIFT and SURF feature descriptors obtains good recognition rates on a well-known actions recognition dataset. Compared with more sophisticated scheme of 3D based HoG/HoF descriptors, proposed method is easier to compute and simpler to understand.

On-Road Car Detection System Using VD-GMM 2.0 (차량검출 GMM 2.0을 적용한 도로 위의 차량 검출 시스템 구축)

  • Lee, Okmin;Won, Insu;Lee, Sangmin;Kwon, Jangwoo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.40 no.11
    • /
    • pp.2291-2297
    • /
    • 2015
  • This paper presents a vehicle detection system using the video as a input image what has moving of vehicles.. Input image has constraints. it has to get fixed view and downward view obliquely from top of the road. Road detection is required to use only the road area in the input image. In introduction, we suggest the experiment result and the critical point of motion history image extraction method, SIFT(Scale_Invariant Feature Transform) algorithm and histogram analysis to detect vehicles. To solve these problem, we propose using applied Gaussian Mixture Model(GMM) that is the Vehicle Detection GMM(VDGMM). In addition, we optimize VDGMM to detect vehicles more and named VDGMM 2.0. In result of experiment, each precision, recall and F1 rate is 9%, 53%, 15% for GMM without road detection and 85%, 77%, 80% for VDGMM2.0 with road detection.

A Prostate Segmentation of TRUS Image using Average Shape Model and SIFT Features (평균 형상 모델과 SIFT 특징을 이용한 TRUS 영상의 전립선 분할)

  • Kim, Sang Bok;Seo, Yeong Geon
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.1 no.3
    • /
    • pp.187-194
    • /
    • 2012
  • Prostate cancer is one of the most frequent cancers in men and is a major cause of mortality in the most of countries. In many diagnostic and treatment procedures for prostate disease, transrectal ultrasound(TRUS) images are being used because the cost is low. But, accurate detection of prostate boundaries is a challenging and difficult task due to weak prostate boundaries, speckle noises and the short range of gray levels. This paper proposes a method for automatic prostate segmentation in TRUS images using its average shape model and invariant features. This approach consists of 4 steps. First, it detects the probe position and the two straight lines connected to the probe using edge distribution. Next, it acquires 3 prostate patches which are in the middle of average model. The patches will be used to compare the features of prostate and nonprostate. Next, it compares and classifies which blocks are similar to 3 representative patches. Last, the boundaries from prior classification and the rough boundaries from first step are used to determine the segmentation. A number of experiments are conducted to validate this method and results showed that this new approach extracted the prostate boundary with less than 7.78% relative to boundary provided manually by experts.

Multi-view Image Generation from Stereoscopic Image Features and the Occlusion Region Extraction (가려짐 영역 검출 및 스테레오 영상 내의 특징들을 이용한 다시점 영상 생성)

  • Lee, Wang-Ro;Ko, Min-Soo;Um, Gi-Mun;Cheong, Won-Sik;Hur, Nam-Ho;Yoo, Ji-Sang
    • Journal of Broadcast Engineering
    • /
    • v.17 no.5
    • /
    • pp.838-850
    • /
    • 2012
  • In this paper, we propose a novel algorithm that generates multi-view images by using various image features obtained from the given stereoscopic images. In the proposed algorithm, we first create an intensity gradient saliency map from the given stereo images. And then we calculate a block-based optical flow that represents the relative movement(disparity) of each block with certain size between left and right images. And we also obtain the disparities of feature points that are extracted by SIFT(scale-invariant We then create a disparity saliency map by combining these extracted disparity features. Disparity saliency map is refined through the occlusion detection and removal of false disparities. Thirdly, we extract straight line segments in order to minimize the distortion of straight lines during the image warping. Finally, we generate multi-view images by grid mesh-based image warping algorithm. Extracted image features are used as constraints during grid mesh-based image warping. The experimental results show that the proposed algorithm performs better than the conventional DIBR algorithm in terms of visual quality.

Invariant Classification and Detection for Cloth Searching (의류 검색용 회전 및 스케일 불변 이미지 분류 및 검색 기술)

  • Hwang, Inseong;Cho, Beobkeun;Jeon, Seungwoo;Choe, Yunsik
    • Journal of Broadcast Engineering
    • /
    • v.19 no.3
    • /
    • pp.396-404
    • /
    • 2014
  • The field of searching clothing, which is very difficult due to the nature of the informal sector, has been in an effort to reduce the recognition error and computational complexity. However, there is no concrete examples of the whole progress of learning and recognizing for cloth, and the related technologies are still showing many limitations. In this paper, the whole process including identifying both the person and cloth in an image and analyzing both its color and texture pattern is specifically shown for classification. Especially, deformable search descriptor, LBPROT_35 is proposed for identifying the pattern of clothing. The proposed method is scale and rotation invariant, so we can obtain even higher detection rate even though the scale and angle of the image changes. In addition, the color classifier with the color space quantization is proposed not to loose color similarity. In simulation, we build database by training a total of 810 images from the clothing images on the internet, and test some of them. As a result, the proposed method shows a good performance as it has 94.4% matching rate while the former Dense-SIFT method has 63.9%.

Image Mosaicking Using Feature Points Based on Color-invariant (칼라 불변 기반의 특징점을 이용한 영상 모자이킹)

  • Kwon, Oh-Seol;Lee, Dong-Chang;Lee, Cheol-Hee;Ha, Yeong-Ho
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.46 no.2
    • /
    • pp.89-98
    • /
    • 2009
  • In the field of computer vision, image mosaicking is a common method for effectively increasing restricted the field of view of a camera by combining a set of separate images into a single seamless image. Image mosaicking based on feature points has recently been a focus of research because of simple estimation for geometric transformation regardless distortions and differences of intensity generating by motion of a camera in consecutive images. Yet, since most feature-point matching algorithms extract feature points using gray values, identifying corresponding points becomes difficult in the case of changing illumination and images with a similar intensity. Accordingly, to solve these problems, this paper proposes a method of image mosaicking based on feature points using color information of images. Essentially, the digital values acquired from a digital color camera are converted to values of a virtual camera with distinct narrow bands. Values based on the surface reflectance and invariant to the chromaticity of various illuminations are then derived from the virtual camera values and defined as color-invariant values invariant to changing illuminations. The validity of these color-invariant values is verified in a test using a Macbeth Color-Checker under simulated illuminations. The test also compares the proposed method using the color-invariant values with the conventional SIFT algorithm. The accuracy of the matching between the feature points extracted using the proposed method is increased, while image mosaicking using color information is also achieved.