• 제목/요약/키워드: SIFT Features

검색결과 115건 처리시간 0.027초

드론 영상을 이용한 특징점 추출 알고리즘 간의 성능 비교 (Performance Comparison and Analysis between Keypoints Extraction Algorithms using Drone Images)

  • 이충호;김의명
    • 한국측량학회지
    • /
    • 제40권2호
    • /
    • pp.79-89
    • /
    • 2022
  • 드론을 이용하여 촬영한 영상은 소규모 지역에 대하여 고품질의 3차원 공간정보를 빠르게 구축할 수 있어 신속한 의사결정이 필요한 분야에 적용되고 있다. 드론 영상을 기반으로 공간정보를 구축하기 위해서는 인접한 드론 영상 간에 특징점 추출하고 영상 매칭을 수행하여 영상 간의 관계를 결정할 필요가 있다. 이에 본 연구에서는 드론을 이용하여 촬영한 주차장과 호수가 공존하는 지역, 건물이 있는 도심 지역, 자연 지형의 들판 지역의 3가지 대상지역을 선정하고 AKAZE (Accelerated-KAZE), BRISK (Binary Robust Invariant Scalable Keypoints), KAZE, ORB(Oriented FAST and Rotated BRIEF), SIFT (Scale Invariant Feature Transform), and SURF (Speeded Up Robust Features) 알고리즘의 성능을 분석하였다. 특징점 추출 알고리즘의 성능은 추출된 특징점의 분포, 매칭점의 분포, 소요시간, 그리고 매칭 정확도를 비교하였다. 주차장과 호수가 공존하는 지역에서는 BRISK 알고리즘의 속도가 신속하였으며, SURF 알고리즘이 특징점과 매칭점의 분포도와 매칭 정확도에서 우수한 성능을 나타내었다. 건물이 있는 도심 지역에서는 AKAZE 알고리즘의 속도가 신속하였으며 SURF 알고리즘이 특징점과 매칭점의 분포도와 매칭 정확도에서 우수한 성능을 나타내었다. 자연 지형의 들판 지역에서는 SURF 알고리즘의 특징점, 매칭점이 드론으로 촬영한 영상 전반적으로 고르게 분포되어 있으나 AKAZE 알고리즘이 가장 높은 매칭 정확도와 신속한 속도를 나타내었다.

스테레오 비전을 이용한 실시간 인간형 로봇 궤적 추출 및 네비게이션 (Real-time Humanoid Robot Trajectory Estimation and Navigation with Stereo Vision)

  • 박지환;조성호
    • 한국정보과학회논문지:소프트웨어및응용
    • /
    • 제37권8호
    • /
    • pp.641-646
    • /
    • 2010
  • 스테레오 카메라를 갖춘 인간형 로봇이 자율적으로 주변 상황을 인지하면서 목적지까지의 경로를 실시간으로 생성 및 수정하는 간단한 알고리즘을 제시한다. 특징점들을 시각적 이미지에서 추출함으로써 주위의 장애물들을 인지한다. 인간형 로봇의 뒤뚱거리는 보행 움직임을 모델링함으로써 로봇의 중심부 기준에서의 실제 경로를 유추하여 계획된 경로와 비교함으로써 시각적 피드백 제어를 구현하고 성공적인 네비게이션을 수행한다. 실제 인간형 로봇의 네비게이션 실험을 통해 제안된 알고리즘의 가능성을 입증한다.

Nearest-Neighbors Based Weighted Method for the BOVW Applied to Image Classification

  • Xu, Mengxi;Sun, Quansen;Lu, Yingshu;Shen, Chenming
    • Journal of Electrical Engineering and Technology
    • /
    • 제10권4호
    • /
    • pp.1877-1885
    • /
    • 2015
  • This paper presents a new Nearest-Neighbors based weighted representation for images and weighted K-Nearest-Neighbors (WKNN) classifier to improve the precision of image classification using the Bag of Visual Words (BOVW) based models. Scale-invariant feature transform (SIFT) features are firstly extracted from images. Then, the K-means++ algorithm is adopted in place of the conventional K-means algorithm to generate a more effective visual dictionary. Furthermore, the histogram of visual words becomes more expressive by utilizing the proposed weighted vector quantization (WVQ). Finally, WKNN classifier is applied to enhance the properties of the classification task between images in which similar levels of background noise are present. Average precision and absolute change degree are calculated to assess the classification performance and the stability of K-means++ algorithm, respectively. Experimental results on three diverse datasets: Caltech-101, Caltech-256 and PASCAL VOC 2011 show that the proposed WVQ method and WKNN method further improve the performance of classification.

Object Classification based on Weakly Supervised E2LSH and Saliency map Weighting

  • Zhao, Yongwei;Li, Bicheng;Liu, Xin;Ke, Shengcai
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제10권1호
    • /
    • pp.364-380
    • /
    • 2016
  • The most popular approach in object classification is based on the bag of visual-words model, which has several fundamental problems that restricting the performance of this method, such as low time efficiency, the synonym and polysemy of visual words, and the lack of spatial information between visual words. In view of this, an object classification based on weakly supervised E2LSH and saliency map weighting is proposed. Firstly, E2LSH (Exact Euclidean Locality Sensitive Hashing) is employed to generate a group of weakly randomized visual dictionary by clustering SIFT features of the training dataset, and the selecting process of hash functions is effectively supervised inspired by the random forest ideas to reduce the randomcity of E2LSH. Secondly, graph-based visual saliency (GBVS) algorithm is applied to detect the saliency map of different images and weight the visual words according to the saliency prior. Finally, saliency map weighted visual language model is carried out to accomplish object classification. Experimental results datasets of Pascal 2007 and Caltech-256 indicate that the distinguishability of objects is effectively improved and our method is superior to the state-of-the-art object classification methods.

맞대기 용접 이음재 인장시험에서 발생한 음향방출 신호의 웨이블릿 변환과 응용 (A Study on the Wavelet Transform of Acoustic Emission Signals Generated from Fusion-Welded Butt Joints in Steel during Tensile Test and its Applications)

  • 이장규;윤종희;우창기;박성완;김봉각;조대희
    • 한국공작기계학회:학술대회논문집
    • /
    • 한국공작기계학회 2005년도 춘계학술대회 논문집
    • /
    • pp.342-348
    • /
    • 2005
  • This study was carried out fusion-welded butt joints in SWS 490A high strength steel subjected to tensile test that load-deflection curve. The windowed or short-time Fourier transform (WFT or SIFT) makes possible for the analysis of non-stationary or transient signals into a joint time-frequency domain and the wavelet transform (WT) is used to decompose the acoustic emission (AE) signal into various discrete series of sequences over different frequency bands. In this paper, for acoustic emission signal analysis to use a continuous wavelet transform, in which the Gabor wavelet base on a Gaussian window function is applied to the time-frequency domain. A wavelet transform is demonstrated and the plots are very powerful in the recognition of the acoustic emission features. As a result, the technique of acoustic emission is ideally suited to study variables which control time and stress dependent fracture or damage process in metallic materials.

  • PDF

모바일 애플리케이션을 위한 특징점 검출 연산자의 비교 분석 (Evaluation of Feature Extraction and Matching Algorithms for the use of Mobile Application)

  • 이용환;김흥준
    • 반도체디스플레이기술학회지
    • /
    • 제14권4호
    • /
    • pp.56-60
    • /
    • 2015
  • Mobile devices like smartphones and tablets are becoming increasingly capable in terms of processing power. Although they are already used in computer vision, no comparable measurement experiments of the popular feature extraction algorithm have been made yet. That is, local feature descriptors are widely used in many computer vision applications, and recently various methods have been proposed. While there are many evaluations have focused on various aspects of local features, matching accuracy, however there are no comparisons considering on speed trade-offs of recent descriptors such as ORB, FAST and BRISK. In this paper, we try to provide a performance evaluation of feature descriptors, and compare their matching precision and speed in KD-Tree setup with efficient computation of Hamming distance. The experimental results show that the recently proposed real valued descriptors such as ORB and FAST outperform state-of-the-art descriptors such SIFT and SURF in both, speed-up efficiency and precision/recall.

A Hybrid Proposed Framework for Object Detection and Classification

  • Aamir, Muhammad;Pu, Yi-Fei;Rahman, Ziaur;Abro, Waheed Ahmed;Naeem, Hamad;Ullah, Farhan;Badr, Aymen Mudheher
    • Journal of Information Processing Systems
    • /
    • 제14권5호
    • /
    • pp.1176-1194
    • /
    • 2018
  • The object classification using the images' contents is a big challenge in computer vision. The superpixels' information can be used to detect and classify objects in an image based on locations. In this paper, we proposed a methodology to detect and classify the image's pixels' locations using enhanced bag of words (BOW). It calculates the initial positions of each segment of an image using superpixels and then ranks it according to the region score. Further, this information is used to extract local and global features using a hybrid approach of Scale Invariant Feature Transform (SIFT) and GIST, respectively. To enhance the classification accuracy, the feature fusion technique is applied to combine local and global features vectors through weight parameter. The support vector machine classifier is a supervised algorithm is used for classification in order to analyze the proposed methodology. The Pascal Visual Object Classes Challenge 2007 (VOC2007) dataset is used in the experiment to test the results. The proposed approach gave the results in high-quality class for independent objects' locations with a mean average best overlap (MABO) of 0.833 at 1,500 locations resulting in a better detection rate. The results are compared with previous approaches and it is proved that it gave the better classification results for the non-rigid classes.

Image Retrieval Method Based on IPDSH and SRIP

  • Zhang, Xu;Guo, Baolong;Yan, Yunyi;Sun, Wei;Yi, Meng
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제8권5호
    • /
    • pp.1676-1689
    • /
    • 2014
  • At present, the Content-Based Image Retrieval (CBIR) system has become a hot research topic in the computer vision field. In the CBIR system, the accurate extractions of low-level features can reduce the gaps between high-level semantics and improve retrieval precision. This paper puts forward a new retrieval method aiming at the problems of high computational complexities and low precision of global feature extraction algorithms. The establishment of the new retrieval method is on the basis of the SIFT and Harris (APISH) algorithm, and the salient region of interest points (SRIP) algorithm to satisfy users' interests in the specific targets of images. In the first place, by using the IPDSH and SRIP algorithms, we tested stable interest points and found salient regions. The interest points in the salient region were named as salient interest points. Secondary, we extracted the pseudo-Zernike moments of the salient interest points' neighborhood as the feature vectors. Finally, we calculated the similarities between query and database images. Finally, We conducted this experiment based on the Caltech-101 database. By studying the experiment, the results have shown that this new retrieval method can decrease the interference of unstable interest points in the regions of non-interests and improve the ratios of accuracy and recall.

SURF와 RANSAC 알고리즘을 이용한 대응점 필터링 적용 파노라마 이미지 처리 (Matching Points Filtering Applied Panorama Image Processing Using SURF and RANSAC Algorithm)

  • 김정호;김대원
    • 전자공학회논문지
    • /
    • 제51권4호
    • /
    • pp.144-159
    • /
    • 2014
  • 다중의 영상을 이용하여 하나의 파노라마 영상을 제작하는 기법은 컴퓨터 비전, 컴퓨터 그래픽스 등과 같은 여러 분야에서 널리 연구되고 있다. 파노라마 영상은 하나의 카메라에서 얻을 수 있는 영상의 한계, 즉 예를 들어 화각, 화질, 정보량 등의 한계를 극복할 수 있는 좋은 방법으로서 가상현실, 로봇비전 등과 같이 광각의 영상이 요구되는 다양한 분야에서 응용될 수 있다. 파노라마 영상은 단일 영상과 비교하여 보다 큰 몰입감을 제공한다는 점에서 큰 의미를 갖는다. 현재 다양한 파노라마 영상 제작 기법들이 존재하지만, 대부분의 기법들이 공통적으로 파노라마 영상을 구성할 때 각 영상에 존재하는 특징점 및 대응점을 검출하는 방식을 사용하고 있다. 또한, 대응점을 이용한 RANSAC(RANdom SAmple Consensus) 알고리즘을 사용, Homography Matrix를 구하여 영상을 변환하는 방법을 사용한다. 본 논문에서 사용한 SURF(Speeded Up Robust Features) 알고리즘은 영상의 특징점을 검출할 때 영상의 흑백정보와 지역 공간 정보를 활용하는데, 영상의 크기 변화와 시점 검출에 강하며 SIFT(Scale Invariant Features Transform) 알고리즘에 비해 속도가 빠르다는 장점이 있어서 널리 사용되고 있다. SURF 알고리즘은 대응점 검출 시 잘못된 대응점을 검출하는 경우가 생긴다는 단점이 존재하는데 이는 RANSAC 알고리즘의 수행속도를 늦추며, 그로인해 CPU 사용 점유율을 높이기도 한다. 대응점 검출 오류는 파노라마 영상의 정확성 및 선명성을 떨어뜨리는 핵심 요인이 된다. 본 논문에서는 이러한 대응점 검출의 오류를 최소화하기 위하여 대응점 좌표 주변 $3{\times}3$ 영역의 RGB값을 사용하여 잘못된 대응점들을 제거하는 중간 필터링 과정을 수행하고, 문제해결을 시도하는 동시에 파노라마 이미지구성 처리 속도 및 CPU 사용 점유율 등의 성능 향상 결과와 추출된 대응점 감소율, 정확도 등과 관련한 분석 및 평가 결과를 제시하였다.

Bag of Visual Words Method based on PLSA and Chi-Square Model for Object Category

  • Zhao, Yongwei;Peng, Tianqiang;Li, Bicheng;Ke, Shengcai
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제9권7호
    • /
    • pp.2633-2648
    • /
    • 2015
  • The problem of visual words' synonymy and ambiguity always exist in the conventional bag of visual words (BoVW) model based object category methods. Besides, the noisy visual words, so-called "visual stop-words" will degrade the semantic resolution of visual dictionary. In view of this, a novel bag of visual words method based on PLSA and chi-square model for object category is proposed. Firstly, Probabilistic Latent Semantic Analysis (PLSA) is used to analyze the semantic co-occurrence probability of visual words, infer the latent semantic topics in images, and get the latent topic distributions induced by the words. Secondly, the KL divergence is adopt to measure the semantic distance between visual words, which can get semantically related homoionym. Then, adaptive soft-assignment strategy is combined to realize the soft mapping between SIFT features and some homoionym. Finally, the chi-square model is introduced to eliminate the "visual stop-words" and reconstruct the visual vocabulary histograms. Moreover, SVM (Support Vector Machine) is applied to accomplish object classification. Experimental results indicated that the synonymy and ambiguity problems of visual words can be overcome effectively. The distinguish ability of visual semantic resolution as well as the object classification performance are substantially boosted compared with the traditional methods.