• Title/Summary/Keyword: feature detector

Search Result 191, Processing Time 0.036 seconds

Improved image alignment algorithm based on projective invariant for aerial video stabilization

  • Yi, Meng;Guo, Bao-Long;Yan, Chun-Man
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.9
    • /
    • pp.3177-3195
    • /
    • 2014
  • In many moving object detection problems of an aerial video, accurate and robust stabilization is of critical importance. In this paper, a novel accurate image alignment algorithm for aerial electronic image stabilization (EIS) is described. The feature points are first selected using optimal derivative filters based Harris detector, which can improve differentiation accuracy and obtain the precise coordinates of feature points. Then we choose the Delaunay Triangulation edges to find the matching pairs between feature points in overlapping images. The most "useful" matching points that belong to the background are used to find the global transformation parameters using the projective invariant. Finally, intentional motion of the camera is accumulated for correction by Sage-Husa adaptive filtering. Experiment results illustrate that the proposed algorithm is applied to the aerial captured video sequences with various dynamic scenes for performance demonstrations.

DCT-Based Images Retrieval for Rotated Images (회전에 견고한 DCT 기반 영상 검색)

  • Kim, Nam-Yee;Song, Ju-Whan;You, Kang-Soo
    • The Journal of the Korea Contents Association
    • /
    • v.11 no.4
    • /
    • pp.67-73
    • /
    • 2011
  • The image retrieval generally shows the same or similar images to a query image as a result. In the case of rotated image, however, its performance tends to be debased significantly. We propose a method to ensure a reliable image retrieval of rotated images as follows; First, to obtain feature points of query/DB images by Harris Corner Detector; and then, utilizing the feature points, to find the object's axis and query/DB images into rotation invariant images with Principal Components Analysis algorithm. We have experimented with 6,000 natural images which are 256 pixels in diameter. They are 1,000 Wang's images and their rotated images by $30^{\circ}$, $45^{\circ}$, $90^{\circ}$, $135^{\circ}$ and $180^{\circ}$. The simulation results show that the proposed method retrieves rotated images more effectively than the conventional method.

Multi-scale face detector using anchor free method

  • Lee, Dong-Ryeol;Kim, Yoon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.25 no.7
    • /
    • pp.47-55
    • /
    • 2020
  • In this paper, we propose one stage multi-scale face detector based Fully Convolution Network using anchor free method. Recently almost all state-of-the-art face detectors which predict location of faces using anchor-based methods rely on pre-defined anchor boxes. However this face detectors need to hyper-parameters and additional computation in training. The key idea of the proposed method is to eliminate hyper-parameters and additional computation using anchor free method. To do this, we apply two ideas. First, by eliminating the pre-defined set of anchor boxes, we avoid the additional computation and hyper-parameters related to anchor boxes. Second, our detector predicts location of faces using multi-feature maps to reduce foreground/background imbalance issue. Through Quantitative evaluation, the performance of the proposed method is evaluated and analyzed. Experimental results on the FDDB dataset demonstrate the effective of our proposed method.

Robust Viewpoint Estimation Algorithm for Moving Parallax Barrier Mobile 3D Display (이동형 패럴랙스 배리어 모바일 3D 디스플레이를 위한 강인한 시청자 시역 위치 추정 알고리즘)

  • Kim, Gi-Seok;Cho, Jae-Soo;Um, Gi-Mun
    • Journal of Broadcast Engineering
    • /
    • v.17 no.5
    • /
    • pp.817-826
    • /
    • 2012
  • This paper presents a robust viewpoint estimation algorithm for Moving Parallax Barrier mobile 3D display in sudden illumination changes. We analyze the previous viewpoint estimation algorithm that consists of the Viola-Jones face detector and the feature tracking by the Optical-Flow. The sudden changes in illumination decreases the performance of the Optical-flow feature tracker. In order to solve the problem, we define a novel performance measure for the Optical-Flow tracker. The overall performance can be increased by the selective adoption of the Viola-Jones detector and the Optical-flow tracker depending on the performance measure. Various experimental results show the effectiveness of the proposed method.

A study on face area detection using face features (얼굴 특징을 이용한 얼굴영역 검출에 관한 연구)

  • Park, Byung-Joon;Kim, Wan-Tae;Kim, Hyun-Sik
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.13 no.3
    • /
    • pp.206-211
    • /
    • 2020
  • It is Face recognition is a very important process in image monitoring and it is a form of biometric technology. The recognition process involves many variables and is highly complex, so the software development has only begun recently with the development of hardware. Face detection technology using the CCTV is a process that precedes face analysis, and it is a technique that detects where the face is in the image. Research in face detection and recognition has been difficult because the human face reacts sensitively to different environmental conditions, such as lighting, color of skin, direction, angle and facial expression. The utility and importance of face recognition technology is coming into the limelight over time, but many aspects are being overlooked in the facial area detection technology that must precede face recognition. The system in this paper can detect tilted faces that cannot be detected by the AdaBoost detector and It could also be used to detect other objects.

Corroded and loosened bolt detection of steel bolted joints based on improved you only look once network and line segment detector

  • Youhao Ni;Jianxiao Mao;Hao Wang;Yuguang Fu;Zhuo Xi
    • Smart Structures and Systems
    • /
    • v.32 no.1
    • /
    • pp.23-35
    • /
    • 2023
  • Steel bolted joint is an important part of steel structure, and its damage directly affects the bearing capacity and durability of steel structure. Currently, the existing research mainly focuses on the identification of corroded bolts and corroded bolts respectively, and there are few studies on multiple states. A detection framework of corroded and loosened bolts is proposed in this study, and the innovations can be summarized as follows: (i) Vision Transformer (ViT) is introduced to replace the third and fourth C3 module of you-only-look-once version 5s (YOLOv5s) algorithm, which increases the attention weights of feature channels and the feature extraction capability. (ii) Three states of the steel bolts are considered, including corroded bolt, bolt missing and clean bolt. (iii) Line segment detector (LSD) is introduced for bolt rotation angle calculation, which realizes bolt looseness detection. The improved YOLOv5s model was validated on the dataset, and the mean average precision (mAP) was increased from 0.902 to 0.952. In terms of a lab-scale joint, the performance of the LSD algorithm and the Hough transform was compared from different perspective angles. The error value of bolt loosening angle of the LSD algorithm is controlled within 1.09%, less than 8.91% of the Hough transform. Furthermore, the proposed framework was applied to fullscale joints of a steel bridge in China. Synthetic images of loosened bolts were successfully identified and the multiple states were well detected. Therefore, the proposed framework can be alternative of monitoring steel bolted joints for management department.

An Integrated Face Detection and Recognition System (통합된 시스템에서의 얼굴검출과 인식기법)

  • 박동희;이규봉;이유홍;나상동;배철수
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2003.05a
    • /
    • pp.165-170
    • /
    • 2003
  • This paper presents an integrated approach to unconstrained face recognition in arbitrary scenes. The front end of the system comprises of a scale and pose tolerant face detector. Scale normalization is achieved through novel combination of a skin color segmentation and log-polar mapping procedure. Principal component analysis is used with the multi-view approach proposed in[10] to handle the pose variations. For a given color input image, the detector encloses a face in a complex scene within a circular boundary and indicates the position of the nose. Next, for recognition, a radial grid mapping centered on the nose yields a feature vector within the circular boundary. As the width of the color segmented region provides an estimated size for the face, the extracted feature vector is scale normalized by the estimated size. The feature vector is input to a trained neural network classifier for face identification. The system was evaluated using a database of 20 person's faces with varying scale and pose obtained on different complex backgrounds. The performance of the face recognizer was also quite good except for sensitivity to small scale face images. The integrated system achieved average recognition rates of 87% to 92%.

  • PDF

An Integrated Face Detection and Recognition System (통합된 시스템에서의 얼굴검출과 인식기법)

  • 박동희;배철수
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.7 no.6
    • /
    • pp.1312-1317
    • /
    • 2003
  • This paper presents an integrated approach to unconstrained face recognition in arbitrary scenes. The front end of the system comprises of a scale and pose tolerant face detector. Scale normalization is achieved through novel combination of a skin color segmentation and log-polar mapping procedure. Principal component analysis is used with the multi-view approach proposed in[10] to handle the pose variations. For a given color input image, the detector encloses a face in a complex scene within a circular boundary and indicates the position of the nose. Next, for recognition, a radial grid mapping centered on the nose yields a feature vector within the circular boundary. As the width of the color segmented region provides an estimated size for the face, the extracted feature vector is scale normalized by the estimated size. The feature vector is input to a trained neural network classifier for face identification. The system was evaluated using a database of 20 person's faces with varying scale and pose obtained on different complex backgrounds. The performance of the face recognizer was also quite good except for sensitivity to small scale face images. The integrated system achieved average recognition rates of 87% to 92%.

Deep Learning-based Gaze Direction Vector Estimation Network Integrated with Eye Landmark Localization (딥 러닝 기반의 눈 랜드마크 위치 검출이 통합된 시선 방향 벡터 추정 네트워크)

  • Joo, Heeyoung;Ko, Min-Soo;Song, Hyok
    • Journal of Broadcast Engineering
    • /
    • v.26 no.6
    • /
    • pp.748-757
    • /
    • 2021
  • In this paper, we propose a gaze estimation network in which eye landmark position detection and gaze direction vector estimation are integrated into one deep learning network. The proposed network uses the Stacked Hourglass Network as a backbone structure and is largely composed of three parts: a landmark detector, a feature map extractor, and a gaze direction estimator. The landmark detector estimates the coordinates of 50 eye landmarks, and the feature map extractor generates a feature map of the eye image for estimating the gaze direction. And the gaze direction estimator estimates the final gaze direction vector by combining each output result. The proposed network was trained using virtual synthetic eye images and landmark coordinate data generated through the UnityEyes dataset, and the MPIIGaze dataset consisting of real human eye images was used for performance evaluation. Through the experiment, the gaze estimation error showed a performance of 3.9, and the estimation speed of the network was 42 FPS (Frames per second).

VILODE : A Real-Time Visual Loop Closure Detector Using Key Frames and Bag of Words (VILODE : 키 프레임 영상과 시각 단어들을 이용한 실시간 시각 루프 결합 탐지기)

  • Kim, Hyesuk;Kim, Incheol
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.4 no.5
    • /
    • pp.225-230
    • /
    • 2015
  • In this paper, we propose an effective real-time visual loop closure detector, VILODE, which makes use of key frames and bag of visual words (BoW) based on SURF feature points. In order to determine whether the camera has re-visited one of the previously visited places, a loop closure detector has to compare an incoming new image with all previous images collected at every visited place. As the camera passes through new places or locations, the amount of images to be compared continues growing. For this reason, it is difficult for a visual loop closure detector to meet both real-time constraint and high detection accuracy. To address the problem, the proposed system adopts an effective key frame selection strategy which selects and compares only distinct meaningful ones from continuously incoming images during navigation, and so it can reduce greatly image comparisons for loop detection. Moreover, in order to improve detection accuracy and efficiency, the system represents each key frame image as a bag of visual words, and maintains indexes for them using DBoW database system. The experiments with TUM benchmark datasets demonstrates high performance of the proposed visual loop closure detector.