• Title/Summary/Keyword: Illumination Invariant Feature

Search Result 32, Processing Time 0.024 seconds

A Robust Hybrid Method for Face Recognition Under Illumination Variation (조명 변이에 강인한 하이브리드 얼굴 인식 방법)

  • Choi, Sang-Il
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.52 no.10
    • /
    • pp.129-136
    • /
    • 2015
  • We propose a hybrid face recognition to deal with illumination variation. For this, we extract discriminant features by using the different illumination invariant feature extraction methods. In order to utilize both advantages of each method, we evaluate the discriminant power of each feature by using the discriminant distance and then construct a composite feature with only the features that contain a large amount of discriminative information. The experimental results for the Multi-PIE, Yale B, AR and yale databases show that the proposed method outperforms an individual illumination invariant feature extraction method for all the databases.

Invariant-Feature Based Object Tracking Using Discrete Dynamic Swarm Optimization

  • Kang, Kyuchang;Bae, Changseok;Moon, Jinyoung;Park, Jongyoul;Chung, Yuk Ying;Sha, Feng;Zhao, Ximeng
    • ETRI Journal
    • /
    • v.39 no.2
    • /
    • pp.151-162
    • /
    • 2017
  • With the remarkable growth in rich media in recent years, people are increasingly exposed to visual information from the environment. Visual information continues to play a vital role in rich media because people's real interests lie in dynamic information. This paper proposes a novel discrete dynamic swarm optimization (DDSO) algorithm for video object tracking using invariant features. The proposed approach is designed to track objects more robustly than other traditional algorithms in terms of illumination changes, background noise, and occlusions. DDSO is integrated with a matching procedure to eliminate inappropriate feature points geographically. The proposed novel fitness function can aid in excluding the influence of some noisy mismatched feature points. The test results showed that our approach can overcome changes in illumination, background noise, and occlusions more effectively than other traditional methods, including color-tracking and invariant feature-tracking methods.

Extended SURF Algorithm with Color Invariant Feature and Global Feature (컬러 불변 특징과 광역 특징을 갖는 확장 SURF(Speeded Up Robust Features) 알고리즘)

  • Yoon, Hyun-Sup;Han, Young-Joon;Hahn, Hern-Soo
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.46 no.6
    • /
    • pp.58-67
    • /
    • 2009
  • A correspondence matching is one of the important tasks in computer vision, and it is not easy to find corresponding points in variable environment where a scale, rotation, view point and illumination are changed. A SURF(Speeded Up Robust Features) algorithm have been widely used to solve the problem of the correspondence matching because it is faster than SIFT(Scale Invariant Feature Transform) with closely maintaining the matching performance. However, because SURF considers only gray image and local geometric information, it is difficult to match corresponding points on the image where similar local patterns are scattered. In order to solve this problem, this paper proposes an extended SURF algorithm that uses the invariant color and global geometric information. The proposed algorithm can improves the matching performance since the color information and global geometric information is used to discriminate similar patterns. In this paper, the superiority of the proposed algorithm is proved by experiments that it is compared with conventional methods on the image where an illumination and a view point are changed and similar patterns exist.

Object Recogniton for Markerless Augmented Reality Embodiment (마커 없는 증강 현실 구현을 위한 물체인식)

  • Paul, Anjan Kumar;Lee, Hyung-Jin;Kim, Young-Bum;Islam, Mohammad Khairul;Baek, Joong-Hwan
    • Journal of Advanced Navigation Technology
    • /
    • v.13 no.1
    • /
    • pp.126-133
    • /
    • 2009
  • In this paper, we propose an object recognition technique for implementing marker less augmented reality. Scale Invariant Feature Transform (SIFT) is used for finding the local features from object images. These features are invariant to scale, rotation, translation, and partially invariant to illumination changes. Extracted Features are distinct and have matched with different image features in the scene. If the trained image is properly matched, then it is expected to find object in scene. In this paper, an object is found from a scene by matching the template images that can be generated from the first frame of the scene. Experimental results of object recognition for 4 kinds of objects showed that the proposed technique has a good performance.

  • PDF

Viewpoint Unconstrained Face Recognition Based on Affine Local Descriptors and Probabilistic Similarity

  • Gao, Yongbin;Lee, Hyo Jong
    • Journal of Information Processing Systems
    • /
    • v.11 no.4
    • /
    • pp.643-654
    • /
    • 2015
  • Face recognition under controlled settings, such as limited viewpoint and illumination change, can achieve good performance nowadays. However, real world application for face recognition is still challenging. In this paper, we propose using the combination of Affine Scale Invariant Feature Transform (SIFT) and Probabilistic Similarity for face recognition under a large viewpoint change. Affine SIFT is an extension of SIFT algorithm to detect affine invariant local descriptors. Affine SIFT generates a series of different viewpoints using affine transformation. In this way, it allows for a viewpoint difference between the gallery face and probe face. However, the human face is not planar as it contains significant 3D depth. Affine SIFT does not work well for significant change in pose. To complement this, we combined it with probabilistic similarity, which gets the log likelihood between the probe and gallery face based on sum of squared difference (SSD) distribution in an offline learning process. Our experiment results show that our framework achieves impressive better recognition accuracy than other algorithms compared on the FERET database.

Implementation of Improved Object Detection and Tracking based on Camshift and SURF for Augmented Reality Service (증강현실 서비스를 위한 Camshift와 SURF를 개선한 객체 검출 및 추적 구현)

  • Lee, Yong-Hwan;Kim, Heung-Jun
    • Journal of the Semiconductor & Display Technology
    • /
    • v.16 no.4
    • /
    • pp.97-102
    • /
    • 2017
  • Object detection and tracking have become one of the most active research areas in the past few years, and play an important role in computer vision applications over our daily life. Many tracking techniques are proposed, and Camshift is an effective algorithm for real time dynamic object tracking, which uses only color features, so that the algorithm is sensitive to illumination and some other environmental elements. This paper presents and implements an effective moving object detection and tracking to reduce the influence of illumination interference, which improve the performance of tracking under similar color background. The implemented prototype system recognizes object using invariant features, and reduces the dimension of feature descriptor to rectify the problems. The experimental result shows that that the system is superior to the existing methods in processing time, and maintains better problem ratios in various environments.

  • PDF

Image Mosaicking Using Feature Points Based on Color-invariant (칼라 불변 기반의 특징점을 이용한 영상 모자이킹)

  • Kwon, Oh-Seol;Lee, Dong-Chang;Lee, Cheol-Hee;Ha, Yeong-Ho
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.46 no.2
    • /
    • pp.89-98
    • /
    • 2009
  • In the field of computer vision, image mosaicking is a common method for effectively increasing restricted the field of view of a camera by combining a set of separate images into a single seamless image. Image mosaicking based on feature points has recently been a focus of research because of simple estimation for geometric transformation regardless distortions and differences of intensity generating by motion of a camera in consecutive images. Yet, since most feature-point matching algorithms extract feature points using gray values, identifying corresponding points becomes difficult in the case of changing illumination and images with a similar intensity. Accordingly, to solve these problems, this paper proposes a method of image mosaicking based on feature points using color information of images. Essentially, the digital values acquired from a digital color camera are converted to values of a virtual camera with distinct narrow bands. Values based on the surface reflectance and invariant to the chromaticity of various illuminations are then derived from the virtual camera values and defined as color-invariant values invariant to changing illuminations. The validity of these color-invariant values is verified in a test using a Macbeth Color-Checker under simulated illuminations. The test also compares the proposed method using the color-invariant values with the conventional SIFT algorithm. The accuracy of the matching between the feature points extracted using the proposed method is increased, while image mosaicking using color information is also achieved.

Affine Local Descriptors for Viewpoint Invariant Face Recognition

  • Gao, Yongbin;Lee, Hyo Jong
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2014.04a
    • /
    • pp.781-784
    • /
    • 2014
  • Face recognition under controlled settings, such as limited viewpoint and illumination change, can achieve good performance nowadays. However, real world application for face recognition is still challenging. In this paper, we use Affine SIFT to detect affine invariant local descriptors for face recognition under large viewpoint change. Affine SIFT is an extension of SIFT algorithm. SIFT algorithm is scale and rotation invariant, which is powerful for small viewpoint changes in face recognition, but it fails when large viewpoint change exists. In our scheme, Affine SIFT is used for both gallery face and probe face, which generates a series of different viewpoints using affine transformation. Therefore, Affine SIFT allows viewpoint difference between gallery face and probe face. Experiment results show our framework achieves better recognition accuracy than SIFT algorithm on FERET database.

Affine Invariant Local Descriptors for Face Recognition (얼굴인식을 위한 어파인 불변 지역 서술자)

  • Gao, Yongbin;Lee, Hyo Jong
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.3 no.9
    • /
    • pp.375-380
    • /
    • 2014
  • Under controlled environment, such as fixed viewpoints or consistent illumination, the performance of face recognition is usually high enough to be acceptable nowadays. Face recognition is, however, a still challenging task in real world. SIFT(Scale Invariant Feature Transformation) algorithm is scale and rotation invariant, which is powerful only in the case of small viewpoint changes. However, it often fails when viewpoint of faces changes in wide range. In this paper, we use Affine SIFT (Scale Invariant Feature Transformation; ASIFT) to detect affine invariant local descriptors for face recognition under wide viewpoint changes. The ASIFT is an extension of SIFT algorithm to solve this weakness. In our scheme, ASIFT is applied only to gallery face, while SIFT algorithm is applied to probe face. ASIFT generates a series of different viewpoints using affine transformation. Therefore, the ASIFT allows viewpoint differences between gallery face and probe face. Experiment results showed our framework achieved higher recognition accuracy than the original SIFT algorithm on FERET database.

Image Similarity Retrieval using an Scale and Rotation Invariant Region Feature (크기 및 회전 불변 영역 특징을 이용한 이미지 유사성 검색)

  • Yu, Seung-Hoon;Kim, Hyun-Soo;Lee, Seok-Lyong;Lim, Myung-Kwan;Kim, Deok-Hwan
    • Journal of KIISE:Databases
    • /
    • v.36 no.6
    • /
    • pp.446-454
    • /
    • 2009
  • Among various region detector and shape feature extraction method, MSER(Maximally Stable Extremal Region) and SIFT and its variant methods are popularly used in computer vision application. However, since SIFT is sensitive to the illumination change and MSER is sensitive to the scale change, it is not easy to apply the image similarity retrieval. In this paper, we present a Scale and Rotation Invariant Region Feature(SRIRF) descriptor using scale pyramid, MSER and affine normalization. The proposed SRIRF method is robust to scale, rotation, illumination change of image since it uses the affine normalization and the scale pyramid. We have tested the SRIRF method on various images. Experimental results demonstrate that the retrieval performance of the SRIRF method is about 20%, 38%, 11%, 24% better than those of traditional SIFT, PCA-SIFT, CE-SIFT and SURF, respectively.