• Title/Summary/Keyword: features-extracting

Search Result 600, Processing Time 0.021 seconds

Speech Emotion Recognition with SVM, KNN and DSVM

  • Hadhami Aouani ;Yassine Ben Ayed
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.8
    • /
    • pp.40-48
    • /
    • 2023
  • Speech Emotions recognition has become the active research theme in speech processing and in applications based on human-machine interaction. In this work, our system is a two-stage approach, namely feature extraction and classification engine. Firstly, two sets of feature are investigated which are: the first one is extracting only 13 Mel-frequency Cepstral Coefficient (MFCC) from emotional speech samples and the second one is applying features fusions between the three features: Zero Crossing Rate (ZCR), Teager Energy Operator (TEO), and Harmonic to Noise Rate (HNR) and MFCC features. Secondly, we use two types of classification techniques which are: the Support Vector Machines (SVM) and the k-Nearest Neighbor (k-NN) to show the performance between them. Besides that, we investigate the importance of the recent advances in machine learning including the deep kernel learning. A large set of experiments are conducted on Surrey Audio-Visual Expressed Emotion (SAVEE) dataset for seven emotions. The results of our experiments showed given good accuracy compared with the previous studies.

Analysis of Deep Learning-Based Lane Detection Models for Autonomous Driving (자율 주행을 위한 심층 학습 기반 차선 인식 모델 분석)

  • Hyunjong Lee;Euihyun Yoon;Jungmin Ha;Jaekoo Lee
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.18 no.5
    • /
    • pp.225-231
    • /
    • 2023
  • With the recent surge in the autonomous driving market, the significance of lane detection technology has escalated. Lane detection plays a pivotal role in autonomous driving systems by identifying lanes to ensure safe vehicle operation. Traditional lane detection models rely on engineers manually extracting lane features from predefined environments. However, real-world road conditions present diverse challenges, hampering the engineers' ability to extract adaptable lane features, resulting in limited performance. Consequently, recent research has focused on developing deep learning based lane detection models to extract lane features directly from data. In this paper, we classify lane detection models into four categories: cluster-based, curve-based, information propagation-based, and anchor-based methods. We conduct an extensive analysis of the strengths and weaknesses of each approach, evaluate the model's performance on an embedded board, and assess their practicality and effectiveness. Based on our findings, we propose future research directions and potential enhancements.

Pedestrian Classification using CNN's Deep Features and Transfer Learning (CNN의 깊은 특징과 전이학습을 사용한 보행자 분류)

  • Chung, Soyoung;Chung, Min Gyo
    • Journal of Internet Computing and Services
    • /
    • v.20 no.4
    • /
    • pp.91-102
    • /
    • 2019
  • In autonomous driving systems, the ability to classify pedestrians in images captured by cameras is very important for pedestrian safety. In the past, after extracting features of pedestrians with HOG(Histogram of Oriented Gradients) or SIFT(Scale-Invariant Feature Transform), people classified them using SVM(Support Vector Machine). However, extracting pedestrian characteristics in such a handcrafted manner has many limitations. Therefore, this paper proposes a method to classify pedestrians reliably and effectively using CNN's(Convolutional Neural Network) deep features and transfer learning. We have experimented with both the fixed feature extractor and the fine-tuning methods, which are two representative transfer learning techniques. Particularly, in the fine-tuning method, we have added a new scheme, called M-Fine(Modified Fine-tuning), which divideslayers into transferred parts and non-transferred parts in three different sizes, and adjusts weights only for layers belonging to non-transferred parts. Experiments on INRIA Person data set with five CNN models(VGGNet, DenseNet, Inception V3, Xception, and MobileNet) showed that CNN's deep features perform better than handcrafted features such as HOG and SIFT, and that the accuracy of Xception (threshold = 0.5) isthe highest at 99.61%. MobileNet, which achieved similar performance to Xception and learned 80% fewer parameters, was the best in terms of efficiency. Among the three transfer learning schemes tested above, the performance of the fine-tuning method was the best. The performance of the M-Fine method was comparable to or slightly lower than that of the fine-tuningmethod, but higher than that of the fixed feature extractor method.

A Study on Face Component Extraction for Automatic Generation of Personal Avatar (개인아바타 자동 생성을 위한 얼굴 구성요소의 추출에 관한 연구)

  • Choi Jae Young;Hwang Seung Ho;Yang Young Kyu;Whangbo Taeg Ken
    • Journal of Internet Computing and Services
    • /
    • v.6 no.4
    • /
    • pp.93-102
    • /
    • 2005
  • In Recent times, Netizens have frequently use virtual character 'Avatar' schemes in order to present their own identity, there is a strong need for avatars to resemble the user. This paper proposes an extraction technique for facial region and features that are used in generating the avatar automatically. For extraction of facial feature component, the method uses ACM and edge information. Also, in the extraction process of facial region, the proposed method reduces the effect of lights and poor image quality on low resolution pictures. this is achieved by using the variation of facial area size which is employed for external energy of ACM. Our experiments show that the success rate of extracting facial regions is $92{\%}$ and accuracy rate of extracting facial feature components is $83.4{\%}$, our results provide good evidence that the suggested method can extract the facial regions and features accurately, moreover this technique can be used in the process of handling features according to the pattern parts of automatic avatar generation system in the near future.

  • PDF

Facial Recognition Algorithm Based on Edge Detection and Discrete Wavelet Transform

  • Chang, Min-Hyuk;Oh, Mi-Suk;Lim, Chun-Hwan;Ahmad, Muhammad-Bilal;Park, Jong-An
    • Transactions on Control, Automation and Systems Engineering
    • /
    • v.3 no.4
    • /
    • pp.283-288
    • /
    • 2001
  • In this paper, we proposed a method for extracting facial characteristics of human being in an image. Given a pair of gray level sample images taken with and without human being, the face of human being is segmented from the image. Noise in the input images is removed with the help of Gaussian filters. Edge maps are found of the two input images. The binary edge differential image is obtained from the difference of the two input edge maps. A mask for face detection is made from the process of erosion followed by dilation on the resulting binary edge differential image. This mask is used to extract the human being from the two input image sequences. Features of face are extracted from the segmented image. An effective recognition system using the discrete wave let transform (DWT) is used for recognition. For extracting the facial features, such as eyebrows, eyes, nose and mouth, edge detector is applied on the segmented face image. The area of eye and the center of face are found from horizontal and vertical components of the edge map of the segmented image. other facial features are obtained from edge information of the image. The characteristic vectors are extrated from DWT of the segmented face image. These characteristic vectors are normalized between +1 and -1, and are used as input vectors for the neural network. Simulation results show recognition rate of 100% on the learned system, and about 92% on the test images.

  • PDF

Adult Image Detection Using Skin Color and Multiple Features (피부색상과 복합 특징을 이용한 유해영상 인식)

  • Jang, Seok-Woo;Choi, Hyung-Il;Kim, Gye-Young
    • Journal of the Korea Society of Computer and Information
    • /
    • v.15 no.12
    • /
    • pp.27-35
    • /
    • 2010
  • Extracting skin color is significant in adult image detection. However, conventional methods still have essential problems in extracting skin color. That is, colors of human skins are basically not the same because of individual skin difference or difference races. Moreover, skin regions of images may not have identical color due to makeup, different cameras used, etc. Therefore, most of the existing methods use predefined skin color models. To resolve these problems, in this paper, we propose a new adult image detection method that robustly segments skin areas with an input image-adapted skin color distribution model, and verifies if the segmented skin regions contain naked bodies by fusing several representative features through a neural network scheme. Experimental results show that our method outperforms others through various experiments. We expect that the suggested method will be useful in many applications such as face detection and objectionable image filtering.

An Artificial Neural Network Learning Fuzzy Membership Functions for Extracting Color Sketch Features (칼라스케치 특징점 추출을 위한 퍼지 멤버쉽 함수의 신경회로망 학습)

  • Cho, Sung-Mok;Cho, Ok-Lae
    • Journal of the Korea Society of Computer and Information
    • /
    • v.11 no.3
    • /
    • pp.11-20
    • /
    • 2006
  • This paper describes the technique which utilizes a fuzzy neural network to sketch feature extraction in digital images. We configure an artificial neural network and make it learn fuzzy membership functions to decide a local threshold applying to sketch feature extraction. To do this. we put the learning data which is membership functions generated based on optimal feature map of a few standard images into the artificial neural network. The proposed technique extracts sketch features in an images very effectively and rapidly because the input fuzzy variable have some desirable characteristics for feature extraction such as dependency of local intensity and excellent performance and the proposed fuzzy neural network is learned from their membership functions, We show that the fuzzy neural network has a good performance in extracting sketch features without human intervention.

  • PDF

Proposal of a Learning Model for Mobile App Malicious Code Analysis (모바일 앱 악성코드 분석을 위한 학습모델 제안)

  • Bae, Se-jin;Choi, Young-ryul;Rhee, Jung-soo;Baik, Nam-kyun
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.455-457
    • /
    • 2021
  • App is used on mobile devices such as smartphones and also has malicious code, which can be divided into normal and malicious depending on the presence or absence of hacking codes. Because there are many kind of malware, it is difficult to detect directly, we propose a method to detect malicious app using AI. Most of the existing methods are to detect malicious app by extracting features from malicious app. However, the number of types have increased exponentially, making it impossible to detect malicious code. Therefore, we would like to propose two more methods besides detecting malicious app by extracting features from most existing malicious app. The first method is to learn normal app to extract normal's features, as opposed to the existing method of learning malicious app and find abnormalities (malicious app). The second one is an 'ensemble technique' that combines the existing method with the first proposal. These two methods need to be studied so that they can be used in future mobile environment.

  • PDF

The Extracting Method of Key-frame Using Color Layout Descriptor (컬러 레이아웃을 이용한 키 프레임 추출 기법)

  • 김소희;김형준;지수영;김회율
    • Proceedings of the IEEK Conference
    • /
    • 2001.06c
    • /
    • pp.213-216
    • /
    • 2001
  • Key frame extraction is an important method of summarizing a long video. This paper propose a technique to automatically extract several key frames representative of its content from video. We use the color layout descriptor to select key frames from video. For selection of key frames, we calculate similarity of color layout features extracted from video, and extract key frames using similarity. An important aspect of our algorithm is that does not assume a fixed number of key frames per video; instead, it selects the number of appropriate key frames of summarizing a long video Experimental results show that our method using color layout descriptor can successfully select several key frames from a video, and we confirmed that the processing speed for extracting key frames from video is considerably fast.

  • PDF

Fast foreground extraction with local Integral Histogram (지역 인테그럴 히스토그램을 사용한 빠르고 강건한 전경 추출 방법)

  • Jang, Dong-Heon;Jin, Xiang-Hua;Kim, Tae-Yong
    • 한국HCI학회:학술대회논문집
    • /
    • 2008.02a
    • /
    • pp.623-628
    • /
    • 2008
  • We present a new method of extracting foreground object from background image for vision-based game interface. Background Subtraction is an important preprocessing step for extracting the features of tracking objects. The image is divided into the cells where the Local Histogram with Gaussian kernel is computed and compared with the corresponding one using Bhattacharyya distance measure. The histogram-based method is partially robust against illumination change, noise and small moving objects in background. We propose a Multi-Scaled Integral Histogram approach for noise suppression and fast computation.

  • PDF