• Title/Summary/Keyword: Acoustic feature extraction

Search Result 56, Processing Time 0.024 seconds

Diagnosis of Transform Aging using Discrete Wavelet Analysis and Neural Network (이산 웨이블렛 분석과 신경망을 이용한 변압기 열화의 전단)

  • 박재준;윤만영;오승헌;김진승;김성홍;백관현;송영철;권동진
    • Proceedings of the Korean Institute of Electrical and Electronic Material Engineers Conference
    • /
    • 2000.07a
    • /
    • pp.645-650
    • /
    • 2000
  • The discrete wavelet transform is utilized as processing of neural network(NN) to identifying aging state of internal partial discharge in transformer. The discrete wavelet transform is used to produce wavelet coefficients which are used for classification. The mean values of the wavelet coefficients are input into an back-propagation neural network. The networks, after training, can decide if the test signals is aging early state or aging last state, or normal state.

  • PDF

A study on Gabor Filter Bank-based Feature Extraction Algorithm for Analysis of Acoustic data of Emergency Rescue (응급구조 음향데이터 분석을 위한 Gabor 필터뱅크 기반의 특징추출 알고리즘에 대한 연구)

  • Hwang, Inyoung;Chang, Joon-Hyuk
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2015.10a
    • /
    • pp.1345-1347
    • /
    • 2015
  • 본 논문에서는 응급상황이 신고되는 상황에서 수보자에게 전달되는 신고자의 주변음향신호로부터 신고자의 주변상황을 추정하기 위하여 음향의 주파수적 특성 및 변화특성의 모델링 성능이 뛰어난 Gabor 필터뱅크 기반의 특징벡터 추출 기술 및 분류 성능이 뛰어난 심화신경망을 도입한다. 제안하는 Gabor 필터뱅크 기반의 특징벡터 추출 기법은 비음성 구간 검출기를 통하여 음성/비음성을 구분한 후에 비음성 구간에서 23차의 Mel-filter bank 계수를 추출한 후에 이로부터 Gabor 필터를 이용하여 주변상황 추정을 위한 특징벡터를 추출하고, 이로부터 학습된 심화신경망을 통하여 신고자의 장소적 정보를 추정한다. 제안된 기법은 여러 가지 시나리오 환경에서 평가되었으며, 우수한 분류성능을 보였다.

A New Temporal Filtering Method for Improved Automatic Lipreading (향상된 자동 독순을 위한 새로운 시간영역 필터링 기법)

  • Lee, Jong-Seok;Park, Cheol-Hoon
    • The KIPS Transactions:PartB
    • /
    • v.15B no.2
    • /
    • pp.123-130
    • /
    • 2008
  • Automatic lipreading is to recognize speech by observing the movement of a speaker's lips. It has received attention recently as a method of complementing performance degradation of acoustic speech recognition in acoustically noisy environments. One of the important issues in automatic lipreading is to define and extract salient features from the recorded images. In this paper, we propose a feature extraction method by using a new filtering technique for obtaining improved recognition performance. The proposed method eliminates frequency components which are too slow or too fast compared to the relevant speech information by applying a band-pass filter to the temporal trajectory of each pixel in the images containing the lip region and, then, features are extracted by principal component analysis. We show that the proposed method produces improved performance in both clean and visually noisy conditions via speaker-independent recognition experiments.

Convolutional neural network based amphibian sound classification using covariance and modulogram (공분산과 모듈로그램을 이용한 콘볼루션 신경망 기반 양서류 울음소리 구별)

  • Ko, Kyungdeuk;Park, Sangwook;Ko, Hanseok
    • The Journal of the Acoustical Society of Korea
    • /
    • v.37 no.1
    • /
    • pp.60-65
    • /
    • 2018
  • In this paper, a covariance matrix and modulogram are proposed for realizing amphibian sound classification using CNN (Convolutional Neural Network). First of all, a database is established by collecting amphibians sounds including endangered species in natural environment. In order to apply the database to CNN, it is necessary to standardize acoustic signals with different lengths. To standardize the acoustic signals, covariance matrix that gives distribution information and modulogram that contains the information about change over time are extracted and used as input to CNN. The experiment is conducted by varying the number of a convolutional layer and a fully-connected layer. For performance assessment, several conventional methods are considered representing various feature extraction and classification approaches. From the results, it is confirmed that convolutional layer has a greater impact on performance than the fully-connected layer. Also, the performance based on CNN shows attaining the highest recognition rate with 99.07 % among the considered methods.

Diagnosis of Valve Internal Leakage for Ship Piping System using Acoustic Emission Signal-based Machine Learning Approach (선박용 밸브의 내부 누설 진단을 위한 음향방출신호의 머신러닝 기법 적용 연구)

  • Lee, Jung-Hyung
    • Journal of the Korean Society of Marine Environment & Safety
    • /
    • v.28 no.1
    • /
    • pp.184-192
    • /
    • 2022
  • Valve internal leakage is caused by damage to the internal parts of the valve, resulting in accidents and shutdowns of the piping system. This study investigated the possibility of a real-time leak detection method using the acoustic emission (AE) signal generated from the piping system during the internal leakage of a butterfly valve. Datasets of raw time-domain AE signals were collected and postprocessed for each operation mode of the valve in a systematic manner to develop a data-driven model for the detection and classification of internal leakage, by applying machine learning algorithms. The aim of this study was to determine whether it is possible to treat leak detection as a classification problem by applying two classification algorithms: support vector machine (SVM) and convolutional neural network (CNN). The results showed different performances for the algorithms and datasets used. The SVM-based binary classification models, based on feature extraction of data, achieved an overall accuracy of 83% to 90%, while in the case of a multiple classification model, the accuracy was reduced to 66%. By contrast, the CNN-based classification model achieved an accuracy of 99.85%, which is superior to those of any other models based on the SVM algorithm. The results revealed that the SVM classification model requires effective feature extraction of the AE signals to improve the accuracy of multi-class classification. Moreover, the CNN-based classification can be a promising approach to detect both leakage and valve opening as long as the performance of the processor does not degrade.

Context Recognition Using Environmental Sound for Client Monitoring System (피보호자 모니터링 시스템을 위한 환경음 기반 상황 인식)

  • Ji, Seung-Eun;Jo, Jun-Yeong;Lee, Chung-Keun;Oh, Siwon;Kim, Wooil
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.19 no.2
    • /
    • pp.343-350
    • /
    • 2015
  • This paper presents a context recognition method using environmental sound signals, which is applied to a mobile-based client monitoring system. Seven acoustic contexts are defined and the corresponding environmental sound signals are obtained for the experiments. To evaluate the performance of the context recognition, MFCC and LPCC method are employed as feature extraction, and statistical pattern recognition method are used employing GMM and HMM as acoustic models, The experimental results show that LPCC and HMM are more effective at improving context recognition accuracy compared to MFCC and GMM respectively. The recognition system using LPCC and HMM obtains 96.03% in recognition accuracy. These results demonstrate that LPCC is effective to represent environmental sounds which contain more various frequency components compared to human speech. They also prove that HMM is more effective to model the time-varying environmental sounds compared to GMM.

Development of Defect Classification Program by Wavelet Transform and Neural Network and Its Application to AE Signal Deu to Welding Defect (웨이블릿 변환과 인공신경망을 이용한 결함분류 프로그램 개발과 용접부 결함 AE 신호에의 적용 연구)

  • Kim, Seong-Hoon;Lee, Kang-Yong
    • Journal of the Korean Society for Nondestructive Testing
    • /
    • v.21 no.1
    • /
    • pp.54-61
    • /
    • 2001
  • A software package to classify acoustic emission (AE) signals using the wavelet transform and the neural network was developed Both of the continuous and the discrete wavelet transforms are considered, and the error back-propagation neural network is adopted as m artificial neural network algorithm. The signals acquired during the 3-point bending test of specimens which have artificial defects on weld zone are used for the classification of the defects. Features are extracted from the time-frequency plane which is the result of the wavelet transform of signals, and the neural network classifier is tamed using the extracted features to classify the signals. It has been shown that the developed software package is useful to classify AE signals. The difference between the classification results by the continuous and the discrete wavelet transforms is also discussed.

  • PDF

Neural Network Based Classification of Time-Varying Signals Distorted by Shallow Water Environment (천해환경에 의해 변형된 시변신호의 신경망을 통한 식별)

  • Na, Young-Nam;Shim, Tae-Bo;Chang, Duck-Hong;Kim, Chun-Duck
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • 1997.06a
    • /
    • pp.27-34
    • /
    • 1997
  • In this study , we tried to test the classification performance of a neural netow and thereby to examine its applicability to the signals distorted by a shallow water einvironment . We conducted an acoustic experiment iin a shallow sea near Pohang, Korea in which water depth is about 60m. The signals, on which the network has been tested, is ilinear frequency modulated ones centered on one of the frequencies, 200, 400, 600 and 800 Hz, each being swept up or down with bandwidth 100Hz. we considered two transforms, STFT(short-time Fourier transform) and PWVD (pseudo Wigner-Ville distribution), form which power spectra were derived. The training signals were simulated using an acoutic model based on the Fourier synthesis scheme. When the network has been trained on the measured signals of center frequency 600Hz,it gave a little better results than that trained onthe simulated . With the center frequencies varied, the overall performance reached over 90% except one case of center frequency 800Hz. With the feature extraction techniques(STFT and PWVD) varied,the network showed performance comparable to each other . In conclusion , the signals which have been simulated with water depth were successully applied to training a neural network, and the trained network performed well in classifying the signals distorted by a surrounding environment and corrupted by noise.

  • PDF

Improvement and Evaluation of the Korean Large Vocabulary Continuous Speech Recognition Platform (ECHOS) (한국어 음성인식 플랫폼(ECHOS)의 개선 및 평가)

  • Kwon, Suk-Bong;Yun, Sung-Rack;Jang, Gyu-Cheol;Kim, Yong-Rae;Kim, Bong-Wan;Kim, Hoi-Rin;Yoo, Chang-Dong;Lee, Yong-Ju;Kwon, Oh-Wook
    • MALSORI
    • /
    • no.59
    • /
    • pp.53-68
    • /
    • 2006
  • We report the evaluation results of the Korean speech recognition platform called ECHOS. The platform has an object-oriented and reusable architecture so that researchers can easily evaluate their own algorithms. The platform has all intrinsic modules to build a large vocabulary speech recognizer: Noise reduction, end-point detection, feature extraction, hidden Markov model (HMM)-based acoustic modeling, cross-word modeling, n-gram language modeling, n-best search, word graph generation, and Korean-specific language processing. The platform supports both lexical search trees and finite-state networks. It performs word-dependent n-best search with bigram in the forward search stage, and rescores the lattice with trigram in the backward stage. In an 8000-word continuous speech recognition task, the platform with a lexical tree increases 40% of word errors but decreases 50% of recognition time compared to the HTK platform with flat lexicon. ECHOS reduces 40% of recognition errors through incorporation of cross-word modeling. With the number of Gaussian mixtures increasing to 16, it yields word accuracy comparable to the previous lexical tree-based platform, Julius.

  • PDF

Condition Monitoring of an LCD Glass Transfer Robot Based on Wavelet Packet Transform and Artificial Neural Network for Abnormal Sound (LCD 라인의 음향 특성신호에 웨이브렛 변환과 인경신경망회로를 적용한 공정로봇의 건정성 감시 연구)

  • Kim, Eui-Youl;Lee, Sang-Kwon;Jang, Ji-Uk
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.36 no.7
    • /
    • pp.813-822
    • /
    • 2012
  • Abnormal operating sounds radiated from a moving transfer robot in LCD (liquid crystal display) product lines have been used for the fault detection line of a robot instead of other source signals such as vibrations, acoustic emissions, and electrical signals. Its advantage as a source signal makes it possible to monitor the status of multiple faults by using only a microphone, despite a relatively low sensitivity. The wavelet packet transform for feature extraction and the artificial neural network for fault classification are employed. It can be observed that the abnormal operating sound is sufficiently useful as a source signal for the fault diagnosis of mechanical components as well as other source signals.