• Title/Summary/Keyword: Acoustic Feature

Search Result 237, Processing Time 0.021 seconds

Pattern Recognition for the Target Signal Using Acoustic Scattering Feature Parameter (표적신호 음향산란 특징파라미터를 이용한 패턴인식에 관한 연구)

  • 주재훈;신기철;김재수
    • The Journal of the Acoustical Society of Korea
    • /
    • v.19 no.4
    • /
    • pp.93-100
    • /
    • 2000
  • Target signal feature parameters are very important to classify target by active sonar. Two highly correlated broad band pulses separated by time T have a time separation pitch(TSP) of 1/T Hz which is equal to the trough-to-trough or peak-to-peak spacing of its spectrum. In this study, TSP informations which represent feature of each target signal were effectively extracted by the FFT. The extracted TSP feature parameters were also applied to the pattern recognition algorithm to classify target and to analyze their properties.

  • PDF

Speech Recognition Performance Improvement using Gamma-tone Feature Extraction Acoustic Model (감마톤 특징 추출 음향 모델을 이용한 음성 인식 성능 향상)

  • Ahn, Chan-Shik;Choi, Ki-Ho
    • Journal of Digital Convergence
    • /
    • v.11 no.7
    • /
    • pp.209-214
    • /
    • 2013
  • Improve the recognition performance of speech recognition systems as a method for recognizing human listening skills were incorporated into the system. In noisy environments by separating the speech signal and noise, select the desired speech signal. but In terms of practical performance of speech recognition systems are factors. According to recognized environmental changes due to noise speech detection is not accurate and learning model does not match. In this paper, to improve the speech recognition feature extraction using gamma tone and learning model using acoustic model was proposed. The proposed method the feature extraction using auditory scene analysis for human auditory perception was reflected In the process of learning models for recognition. For performance evaluation in noisy environments, -10dB, -5dB noise in the signal was performed to remove 3.12dB, 2.04dB SNR improvement in performance was confirmed.

An Acoustic Investigation of Post-Obstruent Tensification Phenomena

  • Ahn, Hyun-Kee
    • Speech Sciences
    • /
    • v.11 no.4
    • /
    • pp.223-232
    • /
    • 2004
  • This study investigated and compared the acoustic characteristics of the Korean stop sound [k'] in three different phonological environments: the tensified lenis stop [k'] as observed in /prek+kaci/, the fortis stop /k'/ as in /pre+k'aci/, and the fortis stop /k'/ following an obstruent as in /prek+k'aci/. The specific research question was whether or not the tensified lenis stop shares all the acoustic features with the other two kinds of fortis stops. The acoustic measures adopted in this study were H1*-H2*, VOT, length of stop closure, and $F_0$. The major findings were that the three stops showed no significant difference in all the acoustic measures except the length of stop closure. The fortis stop /k'/ following an obstruent showed significantly longer duration of stop closure than the other two stops, both of which showed no significant difference. Based on these phonetic results, this study argued that, for the proper phonological description of post-obstruent tensification, the phonological feature [slack vocal folds] of a lenis stop should be changed into [stiff vocal folds, constricted glottis] that the fortis stops should have.

  • PDF

Emotion Recognition and Expression Method using Bi-Modal Sensor Fusion Algorithm (다중 센서 융합 알고리즘을 이용한 감정인식 및 표현기법)

  • Joo, Jong-Tae;Jang, In-Hun;Yang, Hyun-Chang;Sim, Kwee-Bo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.13 no.8
    • /
    • pp.754-759
    • /
    • 2007
  • In this paper, we proposed the Bi-Modal Sensor Fusion Algorithm which is the emotional recognition method that be able to classify 4 emotions (Happy, Sad, Angry, Surprise) by using facial image and speech signal together. We extract the feature vectors from speech signal using acoustic feature without language feature and classify emotional pattern using Neural-Network. We also make the feature selection of mouth, eyes and eyebrows from facial image. and extracted feature vectors that apply to Principal Component Analysis(PCA) remakes low dimension feature vector. So we proposed method to fused into result value of emotion recognition by using facial image and speech.

A Study on Diagnosis of Transformers Aging Sate Using Wavelet Transform and Neural Network (이산웨이블렛 변환과 신경망을 이용한 변압기 열화상태 진단에 관한 연구)

  • 박재준;송영철;전병훈
    • Journal of the Korean Institute of Electrical and Electronic Material Engineers
    • /
    • v.14 no.1
    • /
    • pp.84-92
    • /
    • 2001
  • In this papers, we proposed the new method in order to diagnosis aging state of transformers. For wavelet transform, Daubechies filter is used, we can obtain wavelet coefficients which is used to extract feature of statistical parameters (maximum value, average value, dispersion skewness, kurtosis) about each acoustic emission signal. Also, these coefficients are used to identify normal and fault signal of internal partial discharge in transformer. As improved method for classification use neural network. Extracted statistical parameters are input into an back-propagation neural network. The number of neurons of hidden layer are obtained through Result of Cross-Validation. The network, after training, can decide whether the test signal is early aging state, alst aging state or normal state. In quantity analysis, capability of proposed method is superior to compared that of classical method.

  • PDF

Condition Monitoring of Check Valve Using Neural Network

  • Lee, Seung-Youn;Jeon, Jeong-Seob;Lyou, Joon
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.2198-2202
    • /
    • 2005
  • In this paper we have presented a condition monitoring method of check valve using neural network. The acoustic emission sensor was used to acquire the condition signals of check valve in direct vessel injection (DVI) test loop. The acquired sensor signal pass through a signal conditioning which are consisted of steps; rejection of background noise, amplification, analogue to digital conversion, extract of feature points. The extracted feature points which represent the condition of check valve was utilized input values of fault diagnosis algorithms using pre-learned neural network. The fault diagnosis algorithm proceeds fault detection, fault isolation and fault identification within limited ranges. The developed algorithm enables timely diagnosis of failure of check valve’s degradation and service aging so that maintenance and replacement could be preformed prior to loss of the safety function. The overall process has been experimented and the results are given to show its effectiveness.

  • PDF

Multi-stage Speech Recognition Using Confidence Vector (신뢰도 벡터 기반의 다단계 음성인식)

  • Jeon, Hyung-Bae;Hwang, Kyu-Woong;Chung, Hoon;Kim, Seung-Hi;Park, Jun;Lee, Yun-Keun
    • MALSORI
    • /
    • no.63
    • /
    • pp.113-124
    • /
    • 2007
  • In this paper, we propose a use of confidence vector as an intermediate input feature for multi-stage based speech recognition architecture to improve recognition accuracy. A multi-stage speech recognition structure is introduced as a method to reduce the computational complexity of the decoding procedure and then accomplish faster speech recognition. Conventional multi-stage speech recognition is usually composed of three stages, acoustic search, lexical search, and acoustic re-scoring. In this paper, we focus on improving the accuracy of the lexical decoding by introducing a confidence vector as an input feature instead of phoneme which was used typically. We take experimental results on 220K Korean Point-of-Interest (POI) domain and the experimental results show that the proposed method contributes on improving accuracy.

  • PDF

F-ratio of Speaker Variability in Emotional Speech

  • Yi, So-Pae
    • Speech Sciences
    • /
    • v.15 no.1
    • /
    • pp.63-72
    • /
    • 2008
  • Various acoustic features were extracted and analyzed to estimate the inter- and intra-speaker variability of emotional speech. Tokens of vowel /a/ from sentences spoken with different modes of emotion (sadness, neutral, happiness, fear and anger) were analyzed. All of the acoustic features (fundamental frequency, spectral slope, HNR, H1-A1 and formant frequency) indicated greater contribution to inter- than intra-speaker variability across all emotions. Each acoustic feature of speech signal showed a different degree of contribution to speaker discrimination in different emotional modes. Sadness and neutral indicated greater speaker discrimination than other emotional modes (happiness, fear, anger in descending order of F-ratio). In other words, the speaker specificity was better represented in sadness and neutral than in happiness, fear and anger with any of the acoustic features.

  • PDF

Performance Improvement of Korean Connected Digit Recognition Based on Acoustic Parameters (음향학적 파라메터를 이용한 한국어 연결숫자인식의 성능개선)

  • 김승희;김형순
    • The Journal of the Acoustical Society of Korea
    • /
    • v.18 no.5
    • /
    • pp.58-62
    • /
    • 1999
  • This paper proposes use of acoustic parameters to improve the discriminability among digit models in Korean connected digit recognition. The proposed method used the logarithmic values of energy ratio between the predetermined frequency bands as additional feature parameters, based on the acoustic-phonetic knowledge. The results of our experiment show that the proposed method reduced the error rate by 46% in comparison with the baseline system. And incorporation of channel compensation technique in the proposed method yielded error reduction of about 69%.

  • PDF

Time-Frequency Analysis of Broadband Acoustic Scattering from Chub Mackerel Scomber japonicus, Goldeye Rockfish Sebastes thompsoni, and Fat Greenling Hexagrammos otakii (고등어(Scomber japonicus), 불볼락(Sebastes thompsoni) 및 쥐노래미(Hexagrammos otakii)에 의한 광대역 음향산란신호의 시간-주파수 분석)

  • Lee, Dae-Jae
    • Korean Journal of Fisheries and Aquatic Sciences
    • /
    • v.48 no.2
    • /
    • pp.221-232
    • /
    • 2015
  • Broadband echoes measured in live chub mackerel Scomber japonicus, goldeye rockfish Sebastes thompsoni, and fat greenling Hexagrammos otakii with different morphologies and internal characteristics were analyzed in time and frequency domains to understand the species-specific echo feature characteristics for classifying fish species. The mean echo image for each time-frequency representation dataset obtained as a function of orientation angle was extracted to mitigate the effect of fish orientation on acoustic scattering. The joint time-frequency content of the broadband echo signals was obtained using the smoothed pseudo-Wigner-Ville distribution (SPWVD). The SPWVDs were analyzed for each echo signature of the three fish species. The results show that the time-frequency analysis provided species-specific echo structure patterns and metrics of the broadband acoustic signals to facilitate fish species classification.