• Title/Summary/Keyword: feature extract

Search Result 1,159, Processing Time 0.034 seconds

Patent Document Similarity Based on Image Analysis Using the SIFT-Algorithm and OCR-Text

  • Park, Jeong Beom;Mandl, Thomas;Kim, Do Wan
    • International Journal of Contents
    • /
    • v.13 no.4
    • /
    • pp.70-79
    • /
    • 2017
  • Images are an important element in patents and many experts use images to analyze a patent or to check differences between patents. However, there is little research on image analysis for patents partly because image processing is an advanced technology and typically patent images consist of visual parts as well as of text and numbers. This study suggests two methods for using image processing; the Scale Invariant Feature Transform(SIFT) algorithm and Optical Character Recognition(OCR). The first method which works with SIFT uses image feature points. Through feature matching, it can be applied to calculate the similarity between documents containing these images. And in the second method, OCR is used to extract text from the images. By using numbers which are extracted from an image, it is possible to extract the corresponding related text within the text passages. Subsequently, document similarity can be calculated based on the extracted text. Through comparing the suggested methods and an existing method based only on text for calculating the similarity, the feasibility is achieved. Additionally, the correlation between both the similarity measures is low which shows that they capture different aspects of the patent content.

LSTM Android Malicious Behavior Analysis Based on Feature Weighting

  • Yang, Qing;Wang, Xiaoliang;Zheng, Jing;Ge, Wenqi;Bai, Ming;Jiang, Frank
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.6
    • /
    • pp.2188-2203
    • /
    • 2021
  • With the rapid development of mobile Internet, smart phones have been widely popularized, among which Android platform dominates. Due to it is open source, malware on the Android platform is rampant. In order to improve the efficiency of malware detection, this paper proposes deep learning Android malicious detection system based on behavior features. First of all, the detection system adopts the static analysis method to extract different types of behavior features from Android applications, and extract sensitive behavior features through Term frequency-inverse Document Frequency algorithm for each extracted behavior feature to construct detection features through unified abstract expression. Secondly, Long Short-Term Memory neural network model is established to select and learn from the extracted attributes and the learned attributes are used to detect Android malicious applications, Analysis and further optimization of the application behavior parameters, so as to build a deep learning Android malicious detection method based on feature analysis. We use different types of features to evaluate our method and compare it with various machine learning-based methods. Study shows that it outperforms most existing machine learning based approaches and detects 95.31% of the malware.

A Video Expression Recognition Method Based on Multi-mode Convolution Neural Network and Multiplicative Feature Fusion

  • Ren, Qun
    • Journal of Information Processing Systems
    • /
    • v.17 no.3
    • /
    • pp.556-570
    • /
    • 2021
  • The existing video expression recognition methods mainly focus on the spatial feature extraction of video expression images, but tend to ignore the dynamic features of video sequences. To solve this problem, a multi-mode convolution neural network method is proposed to effectively improve the performance of facial expression recognition in video. Firstly, OpenFace 2.0 is used to detect face images in video, and two deep convolution neural networks are used to extract spatiotemporal expression features. Furthermore, spatial convolution neural network is used to extract the spatial information features of each static expression image, and the dynamic information feature is extracted from the optical flow information of multiple expression images based on temporal convolution neural network. Then, the spatiotemporal features learned by the two deep convolution neural networks are fused by multiplication. Finally, the fused features are input into support vector machine to realize the facial expression classification. Experimental results show that the recognition accuracy of the proposed method can reach 64.57% and 60.89%, respectively on RML and Baum-ls datasets. It is better than that of other contrast methods.

Damage detection of bridges based on spectral sub-band features and hybrid modeling of PCA and KPCA methods

  • Bisheh, Hossein Babajanian;Amiri, Gholamreza Ghodrati
    • Structural Monitoring and Maintenance
    • /
    • v.9 no.2
    • /
    • pp.179-200
    • /
    • 2022
  • This paper proposes a data-driven methodology for online early damage identification under changing environmental conditions. The proposed method relies on two data analysis methods: feature-based method and hybrid principal component analysis (PCA) and kernel PCA to separate damage from environmental influences. First, spectral sub-band features, namely, spectral sub-band centroids (SSCs) and log spectral sub-band energies (LSSEs), are proposed as damage-sensitive features to extract damage information from measured structural responses. Second, hybrid modeling by integrating PCA and kernel PCA is performed on the spectral sub-band feature matrix for data normalization to extract both linear and nonlinear features for nonlinear procedure monitoring. After feature normalization, suppressing environmental effects, the control charts (Hotelling T2 and SPE statistics) is implemented to novelty detection and distinguish damage in structures. The hybrid PCA-KPCA technique is compared to KPCA by applying support vector machine (SVM) to evaluate the effectiveness of its performance in detecting damage. The proposed method is verified through numerical and full-scale studies (a Bridge Health Monitoring (BHM) Benchmark Problem and a cable-stayed bridge in China). The results demonstrate that the proposed method can detect the structural damage accurately and reduce false alarms by suppressing the effects and interference of environmental variations.

A Facial Feature Area Extraction Method for Improving Face Recognition Rate in Camera Image (일반 카메라 영상에서의 얼굴 인식률 향상을 위한 얼굴 특징 영역 추출 방법)

  • Kim, Seong-Hoon;Han, Gi-Tae
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.5
    • /
    • pp.251-260
    • /
    • 2016
  • Face recognition is a technology to extract feature from a facial image, learn the features through various algorithms, and recognize a person by comparing the learned data with feature of a new facial image. Especially, in order to improve the rate of face recognition, face recognition requires various processing methods. In the training stage of face recognition, feature should be extracted from a facial image. As for the existing method of extracting facial feature, linear discriminant analysis (LDA) is being mainly used. The LDA method is to express a facial image with dots on the high-dimensional space, and extract facial feature to distinguish a person by analyzing the class information and the distribution of dots. As the position of a dot is determined by pixel values of a facial image on the high-dimensional space, if unnecessary areas or frequently changing areas are included on a facial image, incorrect facial feature could be extracted by LDA. Especially, if a camera image is used for face recognition, the size of a face could vary with the distance between the face and the camera, deteriorating the rate of face recognition. Thus, in order to solve this problem, this paper detected a facial area by using a camera, removed unnecessary areas using the facial feature area calculated via a Gabor filter, and normalized the size of the facial area. Facial feature were extracted through LDA using the normalized facial image and were learned through the artificial neural network for face recognition. As a result, it was possible to improve the rate of face recognition by approx. 13% compared to the existing face recognition method including unnecessary areas.

Discriminative Power Feature Selection Method for Motor Imagery EEG Classification in Brain Computer Interface Systems

  • Yu, XinYang;Park, Seung-Min;Ko, Kwang-Eun;Sim, Kwee-Bo
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.13 no.1
    • /
    • pp.12-18
    • /
    • 2013
  • Motor imagery classification in electroencephalography (EEG)-based brain-computer interface (BCI) systems is an important research area. To simplify the complexity of the classification, selected power bands and electrode channels have been widely used to extract and select features from raw EEG signals, but there is still a loss in classification accuracy in the state-of- the-art approaches. To solve this problem, we propose a discriminative feature extraction algorithm based on power bands with principle component analysis (PCA). First, the raw EEG signals from the motor cortex area were filtered using a bandpass filter with ${\mu}$ and ${\beta}$ bands. This research considered the power bands within a 0.4 second epoch to select the optimal feature space region. Next, the total feature dimensions were reduced by PCA and transformed into a final feature vector set. The selected features were classified by applying a support vector machine (SVM). The proposed method was compared with a state-of-art power band feature and shown to improve classification accuracy.

Face Feature Extraction Method ThroughStereo Image's Matching Value (스테레오 영상의 정합값을 통한 얼굴특징 추출 방법)

  • Kim, Sang-Myung;Park, Chang-Han;Namkung, Jae-Chan
    • Journal of Korea Multimedia Society
    • /
    • v.8 no.4
    • /
    • pp.461-472
    • /
    • 2005
  • In this paper, we propose face feature extraction algorithm through stereo image's matching value. The proposed algorithm detected face region by change the RGB color space of skin color information to the YCbCr color space. Applying eye-template from extracted face region geometrical feature vector of feature about distance and lean, nose and mouth between eye extracted. And, Proposed method could do feature of eyes, nose and mouth through stereo image's matching as well as 2D feature information extract. In the experiment, the proposed algorithm shows the consistency rate of 73% in distance within about 1m and the consistency rate of 52%in distance since about 1m.

  • PDF

Pedestrian recognition using differential Haar-like feature based on Adaboost algorithm to apply intelligence wheelchair (지능형 휠체어 적용을 위해 Haar-like의 기울기 특징을 이용한 아다부스트 알고리즘 기반의 보행자 인식)

  • Lee, Sang-Hun;Park, Sang-Hee;Lee, Yeung-Hak;Seo, Hee-Don
    • Journal of Biomedical Engineering Research
    • /
    • v.31 no.6
    • /
    • pp.481-486
    • /
    • 2010
  • In this paper, we suggest an advanced algorithm, to recognize pedestrian/non-pedestrian using differential haar-like feature, which applies Adaboost algorithm to make a strong classification from weak classifications. First, we extract two feature vectors: horizontal haar-like feature and vertical haar-like feature. For the next, we calculate the proposed feature vector using differential haar-like method. And then, a strong classification needs to be obtained from weak classifications for composite recognition method using the differential area of horizontal and vertical haar-like. In the proposed method, we use one feature vector and one strong classification for the first stage of recognition. Based on our experiment, the proposed algorithm shows higher recognition rate compared to the traditional method for the pedestrian and non-pedestrian.

A Robust Hybrid Method for Face Recognition Under Illumination Variation (조명 변이에 강인한 하이브리드 얼굴 인식 방법)

  • Choi, Sang-Il
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.52 no.10
    • /
    • pp.129-136
    • /
    • 2015
  • We propose a hybrid face recognition to deal with illumination variation. For this, we extract discriminant features by using the different illumination invariant feature extraction methods. In order to utilize both advantages of each method, we evaluate the discriminant power of each feature by using the discriminant distance and then construct a composite feature with only the features that contain a large amount of discriminative information. The experimental results for the Multi-PIE, Yale B, AR and yale databases show that the proposed method outperforms an individual illumination invariant feature extraction method for all the databases.

Orthonormal Polynomial based Optimal EEG Feature Extraction for Motor Imagery Brain-Computer Interface

  • Chum, Pharino;Park, Seung-Min;Ko, Kwang-Eun;Sim, Kwee-Bo
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.22 no.6
    • /
    • pp.793-798
    • /
    • 2012
  • In this paper, we explored the new method for extracting feature from the electroencephalography (EEG) signal based on linear regression technique with the orthonormal polynomial bases. At first, EEG signals from electrodes around motor cortex were selected and were filtered in both spatial and temporal filter using band pass filter for alpha and beta rhymic band which considered related to the synchronization and desynchonization of firing neurons population during motor imagery task. Signal from epoch length 1s were fitted into linear regression with Legendre polynomials bases and extract the linear regression weight as final features. We compared our feature to the state of art feature, power band feature in binary classification using support vector machine (SVM) with 5-fold cross validations for comparing the classification accuracy. The result showed that our proposed method improved the classification accuracy 5.44% in average of all subject over power band features in individual subject study and 84.5% of classification accuracy with forward feature selection improvement.