• Title/Summary/Keyword: Feature Point Matching

Search Result 195, Processing Time 0.021 seconds

Development of An Inspection Method for Defect Detection on the Surface of Automotive Parts (자동차 부품 형상 결함 탐지를 위한 측정 방법 개발)

  • Park, Hong-Seok;Tuladhar, Upendra Mani;Shin, Seung-Cheol
    • Journal of the Korean Society of Manufacturing Technology Engineers
    • /
    • v.22 no.3
    • /
    • pp.452-458
    • /
    • 2013
  • Over the past several years, many studies have been carried out in the field of 3D data inspection systems. Several attempts have been made to improve the quality of manufactured parts. The introduction of laser sensors for inspection has made it possible to acquire data at a remarkably high speed. In this paper, a robust inspection technique for detecting defects in 3D pressed parts using laser-scanned data is proposed. Point cloud data are segmented for the extraction of features. These segmented features are used for shape matching during the localization process. An iterative closest point (ICP) algorithm is used for the localization of the scanned model and CAD model. To achieve a higher accuracy rate, the ICP algorithm is modified and then used for matching. To enhance the speed of the matching process, aKd-tree algorithm is used. Then, the deviation of the scanned points from the CAD model is computed.

A NEW LANDSAT IMAGE CO-REGISTRATION AND OUTLIER REMOVAL TECHNIQUES

  • Kim, Jong-Hong;Heo, Joon;Sohn, Hong-Gyoo
    • Proceedings of the KSRS Conference
    • /
    • v.2
    • /
    • pp.594-597
    • /
    • 2006
  • Image co-registration is the process of overlaying two images of the same scene. One of which is a reference image, while the other (sensed image) is geometrically transformed to the one. Numerous methods were developed for the automated image co-registration and it is known as a time-consuming and/or computation-intensive procedure. In order to improve efficiency and effectiveness of the co-registration of satellite imagery, this paper proposes a pre-qualified area matching, which is composed of feature extraction with Laplacian filter and area matching algorithm using correlation coefficient. Moreover, to improve the accuracy of co-registration, the outliers in the initial matching point should be removed. For this, two outlier detection techniques of studentized residual and modified RANSAC algorithm are used in this study. Three pairs of Landsat images were used for performance test, and the results were compared and evaluated in terms of robustness and efficiency.

  • PDF

A New Landsat Image Co-Registration and Outlier Removal Techniques

  • Kim, Jong-Hong;Heo, Joon;Sohn, Hong-Gyoo
    • Korean Journal of Remote Sensing
    • /
    • v.22 no.5
    • /
    • pp.439-443
    • /
    • 2006
  • Image co-registration is the process of overlaying two images of the same scene. One of which is a reference image, while the other (sensed image) is geometrically transformed to the one. Numerous methods were developed for the automated image co-registration and it is known as a timeconsuming and/or computation-intensive procedure. In order to improve efficiency and effectiveness of the co-registration of satellite imagery, this paper proposes a pre-qualified area matching, which is composed of feature extraction with Laplacian filter and area matching algorithm using correlation coefficient. Moreover, to improve the accuracy of co-registration, the outliers in the initial matching point should be removed. For this, two outlier detection techniques of studentized residual and modified RANSAC algorithm are used in this study. Three pairs of Landsat images were used for performance test, and the results were compared and evaluated in terms of robustness and efficiency.

Study on a Robust Object Tracking Algorithm Based on Improved SURF Method with CamShift

  • Ahn, Hyochang;Shin, In-Kyoung
    • Journal of the Korea Society of Computer and Information
    • /
    • v.23 no.1
    • /
    • pp.41-48
    • /
    • 2018
  • Recently, surveillance systems are widely used, and one of the key technologies in this surveillance system is to recognize and track objects. In order to track a moving object robustly and efficiently in a complex environment, it is necessary to extract the feature points in the interesting object and to track the object using the feature points. In this paper, we propose a method to track interesting objects in real time by eliminating unnecessary information from objects, generating feature point descriptors using only key feature points, and reducing computational complexity for object recognition. Experimental results show that the proposed method is faster and more robust than conventional methods, and can accurately track objects in various environments.

Keyword Spotting on Hangul Document Images Using Character Feature Models (문자 별 특징 모델을 이용한 한글 문서 영상에서 키워드 검색)

  • Park, Sang-Cheol;Kim, Soo-Hyung;Choi, Deok-Jai
    • The KIPS Transactions:PartB
    • /
    • v.12B no.5 s.101
    • /
    • pp.521-526
    • /
    • 2005
  • In this Paper, we propose a keyword spotting system as an alternative to searching system for poor quality Korean document images and compare the Proposed system with an OCR-based document retrieval system. The system is composed of character segmentation, feature extraction for the query keyword, and word-to-word matching. In the character segmentation step, we propose an effective method to remove the connectivity between adjacent characters and a character segmentation method by making the variance of character widths minimum. In the query creation step, feature vector for the query is constructed by a combination of a character model by typeface. In the matching step, word-to-word matching is applied base on a character-to-character matching. We demonstrated that the proposed keyword spotting system is more efficient than the OCR-based one to search a keyword on the Korean document images, especially when the quality of documents is quite poor and point size is small.

Study on Three-dimension Reconstruction to Low Resolution Image of Crops (작물의 저해상도 이미지에 대한 3차원 복원에 관한 연구)

  • Oh, Jang-Seok;Hong, Hyung-Gil;Yun, Hae-Yong;Cho, Yong-Jun;Woo, Seong-Yong;Song, Su-Hwan;Seo, Kap-Ho;Kim, Dae-Hee
    • Journal of the Korean Society of Manufacturing Process Engineers
    • /
    • v.18 no.8
    • /
    • pp.98-103
    • /
    • 2019
  • A more accurate method of feature point extraction and matching for three-dimensional reconstruction using low-resolution images of crops is proposed herein. This method is important in basic computer vision. In addition to three-dimensional reconstruction from exact matching, map-making and camera location information such as simultaneous localization and mapping can be calculated. The results of this study suggest applicable methods for low-resolution images that produce accurate results. This is expected to contribute to a system that measures crop growth condition.

Hand Gesture Recognition for Understanding Conducting Action (지휘행동 이해를 위한 손동작 인식)

  • Je, Hong-Mo;Kim, Ji-Man;Kim, Dai-Jin
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2007.10c
    • /
    • pp.263-266
    • /
    • 2007
  • We introduce a vision-based hand gesture recognition fer understanding musical time and patterns without extra special devices. We suggest a simple and reliable vision-based hand gesture recognition having two features First, the motion-direction code is proposed, which is a quantized code for motion directions. Second, the conducting feature point (CFP) where the point of sudden motion changes is also proposed. The proposed hand gesture recognition system extracts the human hand region by segmenting the depth information generated by stereo matching of image sequences. And then, it follows the motion of the center of the gravity(COG) of the extracted hand region and generates the gesture features such as CFP and the direction-code finally, we obtain the current timing pattern of beat and tempo of the playing music. The experimental results on the test data set show that the musical time pattern and tempo recognition rate is over 86.42% for the motion histogram matching, and 79.75% fer the CFP tracking only.

  • PDF

Confidence Measure of Depth Map for Outdoor RGB+D Database (야외 RGB+D 데이터베이스 구축을 위한 깊이 영상 신뢰도 측정 기법)

  • Park, Jaekwang;Kim, Sunok;Sohn, Kwanghoon;Min, Dongbo
    • Journal of Korea Multimedia Society
    • /
    • v.19 no.9
    • /
    • pp.1647-1658
    • /
    • 2016
  • RGB+D database has been widely used in object recognition, object tracking, robot control, to name a few. While rapid advance of active depth sensing technologies allows for the widespread of indoor RGB+D databases, there are only few outdoor RGB+D databases largely due to an inherent limitation of active depth cameras. In this paper, we propose a novel method used to build outdoor RGB+D databases. Instead of using active depth cameras such as Kinect or LIDAR, we acquire a pair of stereo image using high-resolution stereo camera and then obtain a depth map by applying stereo matching algorithm. To deal with estimation errors that inevitably exist in the depth map obtained from stereo matching methods, we develop an approach that estimates confidence of depth maps based on unsupervised learning. Unlike existing confidence estimation approaches, we explicitly consider a spatial correlation that may exist in the confidence map. Specifically, we focus on refining confidence feature with the assumption that the confidence feature and resultant confidence map are smoothly-varying in spatial domain and are highly correlated to each other. Experimental result shows that the proposed method outperforms existing confidence measure based approaches in various benchmark dataset.

Orientation Analysis between UAV Video and Photos for 3D Measurement of Bridges (교량의 3차원 측정을 위한 UAV 비디오와 사진의 표정 분석)

  • Han, Dongyeob;Park, Jae Bong;Huh, Jungwon
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.36 no.6
    • /
    • pp.451-456
    • /
    • 2018
  • UAVs (Unmanned Aerial Vehicles) are widely used for maintenance and monitoring of facilities. It is necessary to acquire a high-resolution image for evaluating the appearance state of the facility in safety inspection. In addition, it is essential to acquire the video data in order to acquire data over a wide area rapidly. In general, since video data does not include position information, it is difficult to analyze the actual size of the inspection object quantitatively. In this study, we evaluated the utilization of 3D point cloud data of bridges using a matching between video frames and reference photos. The drones were used to acquire video and photographs. And exterior orientations of the video frames were generated through feature point matching with reference photos. Experimental results showed that the accuracy of the video frame data is similar to that of the reference photos. Furthermore, the point cloud data generated by using video frames represented the shape and size of bridges with usable accuracy. If the stability of the product is verified through the matching test of various conditions in the future, it is expected that the video-based facility modeling and inspection will be effectively conducted.

A Study on the Visual Odometer using Ground Feature Point (지면 특징점을 이용한 영상 주행기록계에 관한 연구)

  • Lee, Yoon-Sub;Noh, Gyung-Gon;Kim, Jin-Geol
    • Journal of the Korean Society for Precision Engineering
    • /
    • v.28 no.3
    • /
    • pp.330-338
    • /
    • 2011
  • Odometry is the critical factor to estimate the location of the robot. In the mobile robot with wheels, odometry can be performed using the information from the encoder. However, the information of location in the encoder is inaccurate because of the errors caused by the wheel's alignment or slip. In general, visual odometer has been used to compensate for the kinetic errors of robot. In case of using the visual odometry under some robot system, the kinetic analysis is required for compensation of errors, which means that the conventional visual odometry cannot be easily applied to the implementation of the other type of the robot system. In this paper, the novel visual odometry, which employs only the single camera toward the ground, is proposed. The camera is mounted at the center of the bottom of the mobile robot. Feature points of the ground image are extracted by using median filter and color contrast filter. In addition, the linear and angular vectors of the mobile robot are calculated with feature points matching, and the visual odometry is performed by using these linear and angular vectors. The proposed odometry is verified through the experimental results of driving tests using the encoder and the new visual odometry.