• Title/Summary/Keyword: Nonlinear speech enhancement

Search Result 12, Processing Time 0.023 seconds

IMM Algorithm with NPHMM for Speech Enhancement (음성 향상을 위한 NPHMM을 갖는 IMM 알고리즘)

  • Lee, Ki-Yong
    • Speech Sciences
    • /
    • v.11 no.4
    • /
    • pp.53-66
    • /
    • 2004
  • The nonlinear speech enhancement method with interactive parallel-extended Kalman filter is applied to speech contaminated by additive white noise. To represent the nonlinear and nonstationary nature of speech. we assume that speech is the output of a nonlinear prediction HMM (NPHMM) combining both neural network and HMM. The NPHMM is a nonlinear autoregressive process whose time-varying parameters are controlled by a hidden Markov chain. The simulation results shows that the proposed method offers better performance gains relative to the previous results [6] with slightly increased complexity.

  • PDF

Nonlinear Speech Enhancement Method for Reducing the Amount of Speech Distortion According to Speech Statistics Model (음성 통계 모형에 따른 음성 왜곡량 감소를 위한 비선형 음성강조법)

  • Choi, Jae-Seung
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.16 no.3
    • /
    • pp.465-470
    • /
    • 2021
  • A robust speech recognition technology is required that does not degrade the performance of speech recognition and the quality of the speech when speech recognition is performed in an actual environment of the speech mixed with noise. With the development of such speech recognition technology, it is necessary to develop an application that achieves stable and high speech recognition rate even in a noisy environment similar to the human speech spectrum. Therefore, this paper proposes a speech enhancement algorithm that processes a noise suppression based on the MMSA-STSA estimation algorithm, which is a short-time spectral amplitude method based on the error of the least mean square. This algorithm is an effective nonlinear speech enhancement algorithm based on a single channel input and has high noise suppression performance. Moreover this algorithm is a technique that reduces the amount of distortion of the speech based on the statistical model of the speech. In this experiment, in order to verify the effectiveness of the MMSA-STSA estimation algorithm, the effectiveness of the proposed algorithm is verified by comparing the input speech waveform and the output speech waveform.

Nonlinear Acoustic Echo Suppressor based on Volterra Filter using Least Squares (Least Squares 기반의 Volterra Filter를 이용한 비선형 반향신호 억제기)

  • Park, Jihwan;Lee, Bong-Ki;Chang, Joon-Hyuk
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.12
    • /
    • pp.205-209
    • /
    • 2013
  • A conventional acoustic echo suppressor (AES) considering only room impulse response between a loudspeaker and a microphone eliminates acoustic echo from the microphone input. However, in a nonlinear acoustic echo environment, the conventional AES degraded because of a nonlinearity of the loudspeaker. In this paper, we adopt AES based on the frequency-domain second-order Volterra filter using Least Square method. For comparing performances, we conduct objective tests including Echo Return Loss Enhancement (ERLE) and Speech Attenuation (SA). The proposed algorithm shows better performance than the conventional in both linear and nonlinear acoustic echo environments.

Effects on the Speech Enhancement Algorithms for Sensorineural Hearing Impairment and Normal Listeners (배경잡음하에서의 감음신경성난청과 정상청력자의 어음인지향상 연구)

  • Kim, D.W.;Kim, I.Y.;Youn, G.W.
    • Proceedings of the KOSOMBE Conference
    • /
    • v.1998 no.11
    • /
    • pp.171-172
    • /
    • 1998
  • Recent development of digital technology has offered new possibilities for noticeable advances of hearing aids. Using the digital technology, it is possible to equip hearing aids with powerful features such as multi-channel nonlinear compression amplification and the feedback cancellation, these are often difficult to implement with analog circuits. Still, speech in noise is one of the major complaints of not only hearing impaired persons but also normal listeners. This paper describes speech intelligibility in background noise for both normal and hearing impaired listeners. Speech enhancement algorithms were implemented and compared for normal and sensorineural hearing impairment listeners.

  • PDF

Multi-channel Speech Enhancement Using Blind Source Separation and Cross-channel Wiener Filtering

  • Jang, Gil-Jin;Choi, Chang-Kyu;Lee, Yong-Beom;Kim, Jeong-Su;Kim, Sang-Ryong
    • The Journal of the Acoustical Society of Korea
    • /
    • v.23 no.2E
    • /
    • pp.56-67
    • /
    • 2004
  • Despite abundant research outcomes of blind source separation (BSS) in many types of simulated environments, their performances are still not satisfactory to be applied to the real environments. The major obstacle may seem the finite filter length of the assumed mixing model and the nonlinear sensor noises. This paper presents a two-step speech enhancement method with multiple microphone inputs. The first step performs a frequency-domain BSS algorithm to produce multiple outputs without any prior knowledge of the mixed source signals. The second step further removes the remaining cross-channel interference by a spectral cancellation approach using a probabilistic source absence/presence detection technique. The desired primary source is detected every frame of the signal, and the secondary source is estimated in the power spectral domain using the other BSS output as a reference interfering source. Then the estimated secondary source is subtracted to reduce the cross-channel interference. Our experimental results show good separation enhancement performances on the real recordings of speech and music signals compared to the conventional BSS methods.

Split Model Speech Analysis Techniques for Speech Signal Enhancement

  • Park, Young-Ho;You, Kwang-Bock;Bae, Myung-Jin
    • Proceedings of the IEEK Conference
    • /
    • 1999.11a
    • /
    • pp.1135-1138
    • /
    • 1999
  • In this paper, The Split Model Analysis Algorithm, which can generate the wideband speech signal from the spectral information of narrowband signal, is developed. The Split Model Analysis Algorithm deals with the separation of the 10$\^$th/ order LPC model into five cascade-connected 2$\^$nd/ order model. The use of the less complex 2$\^$nd/ order models allows for the exclusion of the complicated nonlinear relationships between model parameters and all the poles of the LPC model. The relationships between the model parameters and its corresponding analog poles is proved and applied to each 2$\^$nd/ order model. The wideband speech signal is obtained by changing only the sampling rate.

  • PDF

Noise-Biased Compensation of Minimum Statistics Method using a Nonlinear Function and A Priori Speech Absence Probability for Speech Enhancement (음질향상을 위해 비선형 함수와 사전 음성부재확률을 이용한 최소통계법의 잡음전력편의 보상방법)

  • Lee, Soo-Jeong;Lee, Gang-Seong;Kim, Sun-Hyob
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.1
    • /
    • pp.77-83
    • /
    • 2009
  • This paper proposes a new noise-biased compensation of minimum statistics(MS) method using a nonlinear function and a priori speech absence probability(SAP) for speech enhancement in non-stationary noisy environments. The minimum statistics(MS) method is well known technique for noise power estimation in non-stationary noisy environments. It tends to bias the noise estimate below that of true noise level. The proposed method is combined with an adaptive parameter based on a sigmoid function and a priori speech absence probability (SAP) for biased compensation. Specifically. we apply the adaptive parameter according to the a posteriori SNR. In addition, when the a priori SAP equals unity, the adaptive biased compensation factor separately increases ${\delta}_{max}$ each frequency bin, and vice versa. We evaluate the estimation of noise power capability in highly non-stationary and various noise environments, the improvement in the segmental signal-to-noise ratio (SNR), and the Itakura-Saito Distortion Measure (ISDM) integrated into a spectral subtraction (SS). The results shows that our proposed method is superior to the conventional MS approach.

A simulation study of speech perception enhancement for cochlear implant patients using companding in noisy environment (잡음 환경에서 압신을 이용한 인공 와우 환자의 언어 인지 향상 시뮬레이션 연구)

  • Lee Young-Woo;Ji Yoon-Sang;Lee Jong-Shil;Kim In-Young;Kim Sun-I.;Hong Sung-Hwa;Lee Sang-Min
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.5 s.311
    • /
    • pp.79-87
    • /
    • 2006
  • In this study, we evaluated the performance of a companding strategy as a preprocessing for speech enhancement and noise reduction. The proposed algorithm is based on two tone suppression that is human's hearing characteristics. This algorithm enhances spectral peak of speech signal and reduces background noise, however it has tradeoff characteristics between speech distortion and noise reduction due to limited channel number and nonlinear block. Therefore, we designed two different companding structures that have relative characteristics of noise reduction and speech distortion and found suitable companding structures by difference of individual speech perception ability in noise environment. Thus we proposed speech perception enhancement of cochlear implant user in noise environment with low SNR. The performance of the proposed algorithm was evaluated through 5 normal hearing listeners using noise band simulation. Improvement of speech perception was observed for all subjects and each subject preferred the different type of companding structure.

An Adaptive Microphone Array with Linear Phase Response (선형 위상 특성을 갖는 적응 마이크로폰 어레이)

  • Kang, Hong-Gu;Youn, Dae-Hui;Cha, Il-Hwan
    • The Journal of the Acoustical Society of Korea
    • /
    • v.11 no.3
    • /
    • pp.53-60
    • /
    • 1992
  • Many adaptive beamforming methods have been studied for interference cancellation and speech signal enhancement in telephone conference and auditorium. Main aspect of adaptive beamforming methods for speech signal processing is different from radar, sonar and seismic signal processing because desire output signal should be apt to the human ear. Considering that phase of speech is quite insensible to the human ear, Sondhi proposed a nonlinear constrained optimization technique whose constraint was on the magnitude transfer function from the source to the output. In real environment the phase response of the speech signal affects the human auditorium system. So it is desirable to design linear phase system. In this paper, linear phase beamformer is proposed and sample processing algorithm is also proposed for real time consideration Simulation results show that the proposed algorithm yields more consistent beam patterns and deep nulls to the noise direction than Sondhi's.

  • PDF