• Title/Summary/Keyword: SIFT features

Search Result 115, Processing Time 0.033 seconds

ISAR Cross-Range Scaling for a Maneuvering Target (기동표적에 대한 ISAR Cross-Range Scaling)

  • Kang, Byung-Soo;Bae, Ji-Hoon;Kim, Kyung-Tae;Yang, Eun-Jung
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • v.25 no.10
    • /
    • pp.1062-1068
    • /
    • 2014
  • In this paper, a novel approach estimating target's rotation velocity(RV) is proposed for inverse synthetic aperture radar(ISAR) cross-range scaling(CRS). Scale invariant feature transform(SIFT) is applied to two sequently generated ISAR images for extracting non-fluctuating scatterers. Considering the fact that the distance between target's rotation center(RC) and SIFT features is same, we can set a criterion for estimating RV. Then, the criterion is optimized through the proposed method based on particle swarm optimization(PSO) combined with exhaustive search method. Simulation results show that the proposed algorithm can precisely estimate RV of a scenario based maneuvering target without RC information. With the use of the estimated RV, ISAR image can be correctly re-scaled along the cross-range direction.

A reliable quasi-dense corresponding points for structure from motion

  • Oh, Jangseok;Hong, Hyunggil;Cho, Yongjun;Yun, Haeyong;Seo, Kap-Ho;Kim, Hochul;Kim, Mingi;Lee, Onseok
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.9
    • /
    • pp.3782-3796
    • /
    • 2020
  • A three-dimensional (3D) reconstruction is an important research area in computer vision. The ability to detect and match features across multiple views of a scene is a critical initial step. The tracking matrix W obtained from a 3D reconstruction can be applied to structure from motion (SFM) algorithms for 3D modeling. We often fail to generate an acceptable number of features when processing face or medical images because such images typically contain large homogeneous regions with minimal variation in intensity. In this study, we seek to locate sufficient matching points not only in general images but also in face and medical images, where it is difficult to determine the feature points. The algorithm is implemented on an adaptive threshold value, a scale invariant feature transform (SIFT), affine SIFT, speeded up robust features (SURF), and affine SURF. By applying the algorithm to face and general images and studying the geometric errors, we can achieve quasi-dense matching points that satisfy well-functioning geometric constraints. We also demonstrate a 3D reconstruction with a respectable performance by applying a column space fitting algorithm, which is an SFM algorithm.

Scene Recognition Using Local and Global Features (지역적, 전역적 특징을 이용한 환경 인식)

  • Kang, San-Deul;Hwang, Joong-Won;Jung, Hee-Chul;Han, Dong-Yoon;Sim, Sung-Dae;Kim, Jun-Mo
    • Journal of the Korea Institute of Military Science and Technology
    • /
    • v.15 no.3
    • /
    • pp.298-305
    • /
    • 2012
  • In this paper, we propose an integrated algorithm for scene recognition, which has been a challenging computer vision problem, with application to mobile robot localization. The proposed scene recognition method utilizes SIFT and visual words as local-level features and GIST as a global-level feature. As local-level and global-level features complement each other, it results in improved performance for scene recognition. This improved algorithm is of low computational complexity and robust to image distortions.

Sensor Fusion-Based Semantic Map Building (센서융합을 통한 시맨틱 지도의 작성)

  • Park, Joong-Tae;Song, Jae-Bok
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.17 no.3
    • /
    • pp.277-282
    • /
    • 2011
  • This paper describes a sensor fusion-based semantic map building which can improve the capabilities of a mobile robot in various domains including localization, path-planning and mapping. To build a semantic map, various environmental information, such as doors and cliff areas, should be extracted autonomously. Therefore, we propose a method to detect doors, cliff areas and robust visual features using a laser scanner and a vision sensor. The GHT (General Hough Transform) based recognition of door handles and the geometrical features of a door are used to detect doors. To detect the cliff area and robust visual features, the tilting laser scanner and SIFT features are used, respectively. The proposed method was verified by various experiments and showed that the robot could build a semantic map autonomously in various indoor environments.

Panoramic Image Composition Algorithm through Scaling and Rotation Invariant Features (크기 및 회전 불변 특징점을 이용한 파노라마 영상 합성 알고리즘)

  • Kwon, Ki-Won;Lee, Hae-Yeoun;Oh, Duk-Hwan
    • The KIPS Transactions:PartB
    • /
    • v.17B no.5
    • /
    • pp.333-344
    • /
    • 2010
  • This paper addresses the way to compose paronamic images from images taken the same objects. With the spread of digital camera, the panoramic image has been studied to generate with its interest. In this paper, we propose a panoramic image generation method using scaling and rotation invariant features. First, feature points are extracted from input images and matched with a RANSAC algorithm. Then, after the perspective model is estimated, the input image is registered with this model. Since the SURF feature extraction algorithm is adapted, the proposed method is robust against geometric distortions such as scaling and rotation. Also, the improvement of computational cost is achieved. In the experiment, the SURF feature in the proposed method is compared with features from Harris corner detector or the SIFT algorithm. The proposed method is tested by generating panoramic images using $640{\times}480$ images. Results show that it takes 0.4 second in average for computation and is more efficient than other schemes.

A panorama image generation method using FAST algorithm (FAST를 이용한 파노라마 영상 생성 방법)

  • Kim, Jong-ho;Ko, Jin-woong;Yoo, Jisang
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.20 no.3
    • /
    • pp.630-638
    • /
    • 2016
  • In this paper, a feature based panorama image generation algorithm using FAST(Features from Accelerated Segment Test) method that is faster than SIFT(Scale Invariant Feature Transform) and SURF(Speeded Up Robust Features) is proposed. Cylindrical projection is performed to generate natural panorama images with numerous images as input. The occurred error can be minimized by applying RANSAC(Random Sample Consensus) for the matching process. When we synthesize numerous images acquired from different camera angles, we use blending techniques to compensate the distortions by the heterogeneity of border line. In that way, we could get more natural synthesized panorama image. The proposed algorithm can generate natural panorama images regardless the order of input images and tilted images. In addition, the image matching can be faster than the conventional method. As a result of the experiments, distortion was corrected and natural panorama image was generated.

Video Representation via Fusion of Static and Motion Features Applied to Human Activity Recognition

  • Arif, Sheeraz;Wang, Jing;Fei, Zesong;Hussain, Fida
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.7
    • /
    • pp.3599-3619
    • /
    • 2019
  • In human activity recognition system both static and motion information play crucial role for efficient and competitive results. Most of the existing methods are insufficient to extract video features and unable to investigate the level of contribution of both (Static and Motion) components. Our work highlights this problem and proposes Static-Motion fused features descriptor (SMFD), which intelligently leverages both static and motion features in the form of descriptor. First, static features are learned by two-stream 3D convolutional neural network. Second, trajectories are extracted by tracking key points and only those trajectories have been selected which are located in central region of the original video frame in order to to reduce irrelevant background trajectories as well computational complexity. Then, shape and motion descriptors are obtained along with key points by using SIFT flow. Next, cholesky transformation is introduced to fuse static and motion feature vectors to guarantee the equal contribution of all descriptors. Finally, Long Short-Term Memory (LSTM) network is utilized to discover long-term temporal dependencies and final prediction. To confirm the effectiveness of the proposed approach, extensive experiments have been conducted on three well-known datasets i.e. UCF101, HMDB51 and YouTube. Findings shows that the resulting recognition system is on par with state-of-the-art methods.

Vision based place recognition using Bayesian inference with feedback of image retrieval

  • Yi, Hu;Lee, Chang-Woo
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2006.11a
    • /
    • pp.19-22
    • /
    • 2006
  • In this paper we present a vision based place recognition method which uses Bayesian method with feed back of image retrieval. Both Bayesian method and image retrieval method are based on interest features that are invariant to many image transformations. The interest features are detected using Harris-Laplacian detector and then descriptors are generated from the image patches centered at the features' position in the same manner of SIFT. The Bayesian method contains two stages: learning and recognition. The image retrieval result is fed back to the Bayesian recognition to achieve robust and confidence. The experimental results show the effectiveness of our method.

  • PDF

Comparison of Image Matching Method for Automatic Matching of High Resolution SAR Imagery (SAR 영상 자동정합을 위한 영상정합기법의 비교연구)

  • Baek, Sang Ho;Hong, Seung Hwan;Yoo, Su Hong;Sohn, Hong Gyoo
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.34 no.5
    • /
    • pp.1639-1644
    • /
    • 2014
  • SAR satellite can acquire clear imagery regardless of weather and the images are widely used for land management, natural hazard monitoring and many other applications. Automatic image matching technique is necessary for management of a huge amount of SAR data. Nevertheless, it is difficult to assure the accuracy of image matching due to the difference of image-capturing attitude and time. In this paper, we compared performances of MI method, FMT method and SIFT method by applying arbitrary displacement and rotation to TerraSAR-X images and changing resolution of the images. As a result, when the features having specific intensity were distributed well in SAR imagery, MI method could assure 0~2 pixels accuracy even if the images were captured in different geometry. But the accuracy of FMT method was significantly poor for the images having different spatial resolutions and the error was represented by tens or hundreds pixels. Moreover, the ratio of corresponding matching points for SIFT method was only 0~17% and it was difficult for SIFT method to apply to SAR images captured in different geometry.

Visual Attention Algorithm for Object Recognition (물체 인식을 위한 시각 주목 알고리즘)

  • Ryu, Gwang-Geun;Lee, Sang-Hoon;Suh, Il-Hong
    • Proceedings of the KIEE Conference
    • /
    • 2006.04a
    • /
    • pp.306-308
    • /
    • 2006
  • We propose an attention based object recognition system, to recognize object fast and robustly. For this we calculate visual stimulus degrees and make saliency maps. Through this map we find a strongly attentive part of image by stimulus degrees, where local features are extracted to recognize objects.

  • PDF