• Title/Summary/Keyword: Image Feature

Search Result 3,587, Processing Time 0.034 seconds

A Post-Verification Method of Near-Duplicate Image Detection using SIFT Descriptor Binarization (SIFT 기술자 이진화를 이용한 근-복사 이미지 검출 후-검증 방법)

  • Lee, Yu Jin;Nang, Jongho
    • Journal of KIISE
    • /
    • v.42 no.6
    • /
    • pp.699-706
    • /
    • 2015
  • In recent years, as near-duplicate image has been increasing explosively by the spread of Internet and image-editing technology that allows easy access to image contents, related research has been done briskly. However, BoF (Bag-of-Feature), the most frequently used method for near-duplicate image detection, can cause problems that distinguish the same features from different features or the different features from same features in the quantization process of approximating a high-level local features to low-level. Therefore, a post-verification method for BoF is required to overcome the limitation of vector quantization. In this paper, we proposed and analyzed the performance of a post-verification method for BoF, which converts SIFT (Scale Invariant Feature Transform) descriptors into 128 bits binary codes and compares binary distance regarding of a short ranked list by BoF using the codes. Through an experiment using 1500 original images, it was shown that the near-duplicate detection accuracy was improved by approximately 4% over the previous BoF method.

Content-based Image Retrieval using Feature Extraction in Wavelet Transform Domain (웨이브릿 변환 영역에서 특징추출을 이용한 내용기반 영상 검색)

  • 최인호;이상훈
    • Journal of Korea Multimedia Society
    • /
    • v.5 no.4
    • /
    • pp.415-425
    • /
    • 2002
  • In this paper, we present a content-based image retrieval method which is based on the feature extraction in the wavelet transform domain. In order to overcome the drawbacks of the feature vector making up methods which use the global wavelet coefficients in subbands, we utilize the energy value of wavelet coefficients, and the shape-based retrieval of objects is processed by moment which is invariant in translation, scaling, rotation of the objects The proposed methods reduce feature vector size, and make progress performance of classification retrieval which provides fast retrievals times. To offer the abilities of region-based image retrieval, we discussed the image segmentation method which can reduce the effect of an irregular light sources. The image segmentation method uses a region-merging, and candidate regions which are merged were selected by the energy values of high frequency bands in discrete wavelet transform. The region-based image retrieval is executed by using the segmented region information, and the images are retrieved by a color, texture, shape feature vector.

  • PDF

3D feature point extraction technique using a mobile device (모바일 디바이스를 이용한 3차원 특징점 추출 기법)

  • Kim, Jin-Kyum;Seo, Young-Ho
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.256-257
    • /
    • 2022
  • In this paper, we introduce a method of extracting three-dimensional feature points through the movement of a single mobile device. Using a monocular camera, a 2D image is acquired according to the camera movement and a baseline is estimated. Perform stereo matching based on feature points. A feature point and a descriptor are acquired, and the feature point is matched. Using the matched feature points, the disparity is calculated and a depth value is generated. The 3D feature point is updated according to the camera movement. Finally, the feature point is reset at the time of scene change by using scene change detection. Through the above process, an average of 73.5% of additional storage space can be secured in the key point database. By applying the algorithm proposed to the depth ground truth value of the TUM Dataset and the RGB image, it was confirmed that the\re was an average distance difference of 26.88mm compared with the 3D feature point result.

  • PDF

A Camera Pose Estimation Method for Rectangle Feature based Visual SLAM (사각형 특징 기반 Visual SLAM을 위한 자세 추정 방법)

  • Lee, Jae-Min;Kim, Gon-Woo
    • The Journal of Korea Robotics Society
    • /
    • v.11 no.1
    • /
    • pp.33-40
    • /
    • 2016
  • In this paper, we propose a method for estimating the pose of the camera using a rectangle feature utilized for the visual SLAM. A warped rectangle feature as a quadrilateral in the image by the perspective transformation is reconstructed by the Coupled Line Camera algorithm. In order to fully reconstruct a rectangle in the real world coordinate, the distance between the features and the camera is needed. The distance in the real world coordinate can be measured by using a stereo camera. Using properties of the line camera, the physical size of the rectangle feature can be induced from the distance. The correspondence between the quadrilateral in the image and the rectangle in the real world coordinate can restore the relative pose between the camera and the feature through obtaining the homography. In order to evaluate the performance, we analyzed the result of proposed method with its reference pose in Gazebo robot simulator.

Quantitative Evaluation of Nonlinear Shape Normalization Methods for the Recognition of Large-Set Handwrittern Characters (대용량 필기체 문자 인식을 위한 비선형 형태 정규화 방법의 정량적 평가)

  • 이성환;박정선
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.30B no.9
    • /
    • pp.84-93
    • /
    • 1993
  • Recently, several nonlinear shape normalization methods have been proposed in order to compensate for the shape distortions in handwritten characters. In this paper, we review these nonlinear shape normalization methods from the two points of view : feature projection and feature density equalization. The former makes feature projection histogram by projecting a certain feature at each point of input image into horizontal-or vertical-axis and the latter equalizes the feature densities of input image by re-sampling the feature projection histogram. A systematic comparison of these methods has been made based on the following criteria: recognition rate, processing speed, computational complexity and measure of variation. Then, we present the result of quantitative evaluation of each method based on these criteria for a large variety of handwritten Hangul syllables.

  • PDF

2D Shape Recognition System Using Fuzzy Weighted Mean by Statistical Information

  • Woo, Young-Woon;Han, Soo-Whan
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2009.01a
    • /
    • pp.49-54
    • /
    • 2009
  • A fuzzy weighted mean method on a 2D shape recognition system is introduced in this paper. The bispectrum based on third order cumulant is applied to the contour sequence of each image for the extraction of a feature vector. This bispectral feature vector, which is invariant to shape translation, rotation and scale, represents a 2D planar image. However, to obtain the best performance, it should be considered certain criterion on the calculation of weights for the fuzzy weighted mean method. Therefore, a new method to calculate weights using means by differences of feature values and their variances with the maximum distance from differences of feature values. is developed. In the experiments, the recognition results with fifteen dimensional bispectral feature vectors, which are extracted from 11.808 aircraft images based on eight different styles of reference images, are compared and analyzed.

  • PDF

A Defocus Technique based Depth from Lens Translation using Sequential SVD Factorization

  • Kim, Jong-Il;Ahn, Hyun-Sik;Jeong, Gu-Min;Kim, Do-Hyun
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.383-388
    • /
    • 2005
  • Depth recovery in robot vision is an essential problem to infer the three dimensional geometry of scenes from a sequence of the two dimensional images. In the past, many studies have been proposed for the depth estimation such as stereopsis, motion parallax and blurring phenomena. Among cues for depth estimation, depth from lens translation is based on shape from motion by using feature points. This approach is derived from the correspondence of feature points detected in images and performs the depth estimation that uses information on the motion of feature points. The approaches using motion vectors suffer from the occlusion or missing part problem, and the image blur is ignored in the feature point detection. This paper presents a novel approach to the defocus technique based depth from lens translation using sequential SVD factorization. Solving such the problems requires modeling of mutual relationship between the light and optics until reaching the image plane. For this mutuality, we first discuss the optical properties of a camera system, because the image blur varies according to camera parameter settings. The camera system accounts for the camera model integrating a thin lens based camera model to explain the light and optical properties and a perspective projection camera model to explain the depth from lens translation. Then, depth from lens translation is proposed to use the feature points detected in edges of the image blur. The feature points contain the depth information derived from an amount of blur of width. The shape and motion can be estimated from the motion of feature points. This method uses the sequential SVD factorization to represent the orthogonal matrices that are singular value decomposition. Some experiments have been performed with a sequence of real and synthetic images comparing the presented method with the depth from lens translation. Experimental results have demonstrated the validity and shown the applicability of the proposed method to the depth estimation.

  • PDF

FIGURE ALPHABET HYPOTHESIS INSPIRED NEURAL NETWORK RECOGNITION MODEL

  • Ohira, Ryoji;Saiki, Kenji;Nagao, Tomoharu
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2009.01a
    • /
    • pp.547-550
    • /
    • 2009
  • The object recognition mechanism of human being is not well understood yet. On research of animal experiment using an ape, however, neurons that respond to simple shape (e.g. circle, triangle, square and so on) were found. And Hypothesis has been set up as human being may recognize object as combination of such simple shapes. That mechanism is called Figure Alphabet Hypothesis, and those simple shapes are called Figure Alphabet. As one way to research object recognition algorithm, we focused attention to this Figure Alphabet Hypothesis. Getting idea from it, we proposed the feature extraction algorithm for object recognition. In this paper, we described recognition of binarized images of multifont alphabet characters by the recognition model which combined three-layered neural network in the feature extraction algorithm. First of all, we calculated the difference between the learning image data set and the template by the feature extraction algorithm. The computed finite difference is a feature quantity of the feature extraction algorithm. We had it input the feature quantity to the neural network model and learn by backpropagation (BP method). We had the recognition model recognize the unknown image data set and found the correct answer rate. To estimate the performance of the contriving recognition model, we had the unknown image data set recognized by a conventional neural network. As a result, the contriving recognition model showed a higher correct answer rate than a conventional neural network model. Therefore the validity of the contriving recognition model could be proved. We'll plan the research a recognition of natural image by the contriving recognition model in the future.

  • PDF

A Real-time Indoor Place Recognition System Using Image Features Detection (영상 특징 검출 기반의 실시간 실내 장소 인식 시스템)

  • Song, Bok-Deuk;Shin, Bum-Joo;Yang, Hwang-Kyu
    • Journal of the Korean Institute of Electrical and Electronic Material Engineers
    • /
    • v.25 no.1
    • /
    • pp.76-83
    • /
    • 2012
  • In a real-time indoor place recognition system using image features detection, specific markers included in input image should be detected exactly and quickly. However because the same markers in image are shown up differently depending to movement, direction and angle of camera, it is required a method to solve such problems. This paper proposes a technique to extract the features of object without regard to change of the object scale. To support real-time operation, it adopts SURF(Speeded up Robust Features) which enables fast feature detection. Another feature of this system is the user mark designation which makes possible for user to designate marks from input image for location detection in advance. Unlike to use hardware marks, the feature above has an advantage that the designated marks can be used without any manipulation to recognize location in input image.

Implementation of an improved real-time object tracking algorithm using brightness feature information and color information of object

  • Kim, Hyung-Hoon;Cho, Jeong-Ran
    • Journal of the Korea Society of Computer and Information
    • /
    • v.22 no.5
    • /
    • pp.21-28
    • /
    • 2017
  • As technology related to digital imaging equipment is developed and generalized, digital imaging system is used for various purposes in fields of society. The object tracking technology from digital image data in real time is one of the core technologies required in various fields such as security system and robot system. Among the existing object tracking technologies, cam shift technology is a technique of tracking an object using color information of an object. Recently, digital image data using infrared camera functions are widely used due to various demands of digital image equipment. However, the existing cam shift method can not track objects in image data without color information. Our proposed tracking algorithm tracks the object by analyzing the color if valid color information exists in the digital image data, otherwise it generates the lightness feature information and tracks the object through it. The brightness feature information is generated from the ratio information of the width and the height of the area divided by the brightness. Experimental results shows that our tracking algorithm can track objects in real time not only in general image data including color information but also in image data captured by an infrared camera.