• Title/Summary/Keyword: Speech Signals

Search Result 499, Processing Time 0.069 seconds

Chaotic Speech Secure Communication Using Self-feedback Masking Techniques (자기피드백 마스킹 기법을 사용한 카오스 음성비화통신)

  • Lee, Ik-Soo;Ryeo, Ji-Hwan
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.13 no.6
    • /
    • pp.698-703
    • /
    • 2003
  • This paper presents analog secure communication system about safe speech transmission using chaotic signals. We applied various conditions that happen in actuality communication environment modifying chaotic synchronization and chaotic communication schemes and analyzed restoration performance of speech signal to computer simulation. In transmitter, we made the chaotic masking signal which is added voice signal to chaotic signal using PC(Pecora & Carroll) and SFB(self-feedback) control techniques and transmitted encryption signal to noisy communication channel And in order to calculate the degree of restoration performance, we proposed the definition of analog average power of recovered error signals in receiver chaotic system. The simulation results show that feedback control techniques can certify that restoration performance is superior to quantitative data than PC method about masking degree, susceptibility of parameters and channel noise. We experimentally computed the table of relation of parameter fluxion to restoration error rate which is applied the encryption key values to the chaotic secure communication.

A Novel Approach to COVID-19 Diagnosis Based on Mel Spectrogram Features and Artificial Intelligence Techniques

  • Alfaidi, Aseel;Alshahrani, Abdullah;Aljohani, Maha
    • International Journal of Computer Science & Network Security
    • /
    • v.22 no.9
    • /
    • pp.195-207
    • /
    • 2022
  • COVID-19 has remained one of the most serious health crises in recent history, resulting in the tragic loss of lives and significant economic impacts on the entire world. The difficulty of controlling COVID-19 poses a threat to the global health sector. Considering that Artificial Intelligence (AI) has contributed to improving research methods and solving problems facing diverse fields of study, AI algorithms have also proven effective in disease detection and early diagnosis. Specifically, acoustic features offer a promising prospect for the early detection of respiratory diseases. Motivated by these observations, this study conceptualized a speech-based diagnostic model to aid in COVID-19 diagnosis. The proposed methodology uses speech signals from confirmed positive and negative cases of COVID-19 to extract features through the pre-trained Visual Geometry Group (VGG-16) model based on Mel spectrogram images. This is used in addition to the K-means algorithm that determines effective features, followed by a Genetic Algorithm-Support Vector Machine (GA-SVM) classifier to classify cases. The experimental findings indicate the proposed methodology's capability to classify COVID-19 and NOT COVID-19 of varying ages and speaking different languages, as demonstrated in the simulations. The proposed methodology depends on deep features, followed by the dimension reduction technique for features to detect COVID-19. As a result, it produces better and more consistent performance than handcrafted features used in previous studies.

Overlap and Add Sinusoidal Synthesis Method of Speech Signal Lising the Damping Harmonic Magnitude Parameter (감쇄(damping) 하모닉 크기 파라미터를 이용한 음성의 중첩합산 정현파 합성 방법)

  • Park, Jong-Bae;Kim, Young-Joon;Lee, In-Sung
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.34 no.3C
    • /
    • pp.251-256
    • /
    • 2009
  • In this paper, we propose a new method with the improved continuity performance of overlap and add speech signal synthesis method using damping harmonic amplitude parameter. The existing method uses the average value of past and current parameters for the sinusoidal amplitude used as the weight of phase error function. But, the proposed method extracts the more accurate sinusoidal amplitude by using a correlation between the original signals and the synthesized signals for the sinusodal amplitude used as the weights. To verify the performance of the proposed method, we observed the average differential error value between the synthesized signals.

SPECTRAL CHARACTERISTICS OF RESONANCE DISORDERS IN SUBMUCOSAL TYPE CLEFT PALATE PATIENTS (점막하 구개열 환자 공명장애의 스펙트럼 특성 연구)

  • Kim, Hyun-Chul;Leem, Dae-Ho;Baek, Jin-A;Shin, Hyo-Keun;Kim, Oh-Hwan;Kim, Hyun-Ki
    • Maxillofacial Plastic and Reconstructive Surgery
    • /
    • v.28 no.4
    • /
    • pp.310-319
    • /
    • 2006
  • Submucosal type cleft palate is subdivision of the cleft palate. It is very difficult to find submucosal cleft, because when we exam submucosal type cleft palate patients, it seems to be normal. But in fact, there are abnormal union of palatal muscles of submucosal type cleft palate patients. Because of late detection, the treatment - for example, the operation or the speech therapy - for the submucosal type cleft palate patient usually becomes late. Some patients visited our hospital due to speech disorder nevertheless normal intraoral appearance. After precise intraoral examination, we found out submucosal cleft palate. We evaluated the speech before and after surgery of these patients. In this study, we want to find the objective characteristics of submucosal type cleft palate patients, comparing with the normal and the complete cleft palate patients. Experimental groups were 10 submucosal type cleft palate patients and 10 complete cleft palate patients who got the operation in our hospital. And, the controls were 10 normal person. The sentence patterns using in this study were simple 5 vowels. Using CSL program we evaluated the Formant, Bandwidth. We analized the spectral characteristics of speech signals of 3 groups, before and after the operation. In most cases, the formant scores were higher in experimental groups (complete cleft palate group and submucosal type cleft palate group) than controls. There were small differences when speeching /a/, /i/, /e/ between experimental groups and control groups, large differences when speeching /o/, /u/. After surgery the formant scores were decreased in experimental groups (complete cleft palate group and submucosal type cleft palate group). In bandwidth scores, there were no significant differences between experimental groups and controls.

16kbps Windeband Sideband Speech Codec (16kbps 광대역 음성 압축기 개발)

  • 박호종;송재종
    • The Journal of the Acoustical Society of Korea
    • /
    • v.21 no.1
    • /
    • pp.5-10
    • /
    • 2002
  • This paper proposes new 16 kbps wideband speech codec with bandwidth of 7 kHz. The proposed codec decomposes the input speech signal into low-band and high-band signals using QMF (Quadrature Mirror Filter), then AMR (Adaptive Multi Rate) speech codec processes the low-band signal and new transform-domain codec based on G.722.1 wideband cosec compresses the high-band signal. The proposed codec allocates different number of bits to each band in an adaptive way according to the property of input signal, which provides better performance than the codec with the fixed bit allocation scheme. In addition, the proposed cosec processes high-band signal using wavelet transform for better performance. The performance of proposed codec is measured in a subjective method. and the simulations with various speech data show that the proposed coders has better performance than G.722 48 kbps SB-ADPCM.

A Study on Pitch Extraction Method using FIR-STREAK Digital Filter (FIR-STREAK 디지털 필터를 사용한 피치추출 방법에 관한 연구)

  • Lee, Si-U
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.1
    • /
    • pp.247-252
    • /
    • 1999
  • In order to realize a speech coding at low bit rates, a pitch information is useful parameter. In case of extracting an average pitch information form continuous speech, the several pitch errors appear in a frame which consonant and vowel are coexistent; in the boundary between adjoining frames and beginning or ending of a sentence. In this paper, I propose an Individual Pitch (IP) extraction method using residual signals of the FIR-STREAK digital filter in order to restrict the pitch extraction errors. This method is based on not averaging pitch intervals in order to accomodate the changes in each pitch interval. As a result, in case of Ip extraction method suing FIR-STREAK digital filter, I can't find the pitch errors in a frame which consonant and vowel are consistent; in the boundary between adjoining frames and beginning or ending of a sentence. This method has the capability of being applied to many fields, such as speech coding, speech analysis, speech synthesis and speech recognition.

  • PDF

Speech Quality Estimation Algorithm using a Harmonic Modeling of Reverberant Signals (반향 음성 신호의 하모닉 모델링을 이용한 음질 예측 알고리즘)

  • Yang, Jae-Mo;Kang, Hong-Goo
    • Journal of Broadcast Engineering
    • /
    • v.18 no.6
    • /
    • pp.919-926
    • /
    • 2013
  • The acoustic signal from a distance sound source in an enclosed space often produces reverberant sound that varies depending on room impulse response. The estimation of the level of reverberation or the quality of the observed signal is important because it provides valuable information on the condition of system operating environment. It is also useful for designing a dereverberation system. This paper proposes a speech quality estimation method based on the harmonicity of received signal, a unique characteristic of voiced speech. At first, we show that the harmonic signal modeling to a reverberant signal is reasonable. Then, the ratio between the harmonically modeled signal and the estimated non-harmonic signal is used as a measure of standard room acoustical parameter, which is related to speech clarity. Experimental results show that the proposed method successfully estimates speech quality when the reverberation time varies from 0.2s to 1.0s. Finally, we confirm the superiority of the proposed method in both background noise and reverberant environments.

Articulation Scores and Confusion Patterns of the 100 Monosyllable Korean Speech Sounds (우리말 100단음절의 명료도와 오청상에 관한 연구)

  • 유방환;김홍기;노관택
    • Proceedings of the KOR-BRONCHOESO Conference
    • /
    • 1972.03a
    • /
    • pp.1.1-1
    • /
    • 1972
  • It is well known that speech signals are the most riliable materials for the hearing test and there are various difficult problems in the selection of these materials. Because of these difficulties, there is not a confirmed test material of Korean speech sound up to date. For the basis of the test materials, author had studied articulation scores and confusion patterns of 100 monosyllable korean speech sounds in normal listners, in normal listners under various noisy (white noise and speech noise) circumstances, and in patients with hearing loss, The results reveal as follows. 1. Except for perceptive deafness with poor articulation score, Confusion was occured among initial consonants, vowels and final consonants respectively according to their distinctive features under above various test conditions. 2. There is remarkable differences in articulation scores between different kindes of noise under some intensity levels.

  • PDF

Speech Enhancement Based on Improved Minima Controlled Recursive Averaging Incorporating GSAP (전역 음성 부재 확률 기반의 향상된 최소값 제어 재귀평균기법을 이용한 음성 향상 기법)

  • Song, Ji-Hyun;Bang, Dong-Hyeouck;Lee, Sang-Min
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.49 no.1
    • /
    • pp.104-111
    • /
    • 2012
  • In this paper, we propose a novel method to improve the performance of the improved minima controlled recursive averaging (IMCRA). From an examination for various noise environment, it is shown that the IMCRA has a fundamental drawback for the noise power estimate at the offset region of continuity speech signals. Espectially, it is difficult to obtain the robust estimates of the noise power in non-stationary noisy environments that is rapidly changed the spectral characteristics such as babble noise. To overcome the drawback, we apply the global speech absence probability (GSAP) conditioned on both a priori SNR and a posteriori SNR to the speech detection algorithm of IMCRA. With the performance criteria of the ITU-T P.862 perceptual evaluation of speech quality (PESQ) and a composite measure test, we show that the proposed algorithm yields better results compared to the conventional IMCRA-based scheme under various noise environments. In particular, in the case of babble 5 dB, the proposed method produced a remarkable improvement compared to the IMCRA ( PESQ = 0.026, composite measure = 0.029 ).

An Alteration Rule of Formant Transition for Improvement of Korean Demisyllable Based Synthesis by Rule (한국어 반음절단위 규칙합성의 개선을 위한 포만트천이의 변경규칙)

  • Lee, Ki-Young;Choi, Chang-Seok
    • The Journal of the Acoustical Society of Korea
    • /
    • v.15 no.4
    • /
    • pp.98-104
    • /
    • 1996
  • This paper propose the alteraton rule to compensate a formant trasition of several connected vowels for improving an unnatural synthesized continuous speech which is concatenated by each demisyllable without coarticulated formant transition for use in dmisyllable based synthesis by rule. To fullfill each formant transition part, the database of 42 stationary vowels which are segmented from the stable part of each vowels is appended to the one of Korean demisyllables, and the resonance circuit used in formant synthesis is employed to change the formant frequency of speech signals. To evaluate the synthesied speech by this rule, we carried out the alteration rule for connected vowels of the synthesized speech based on demisyllable, and compare spectrogram and MOS tested scores with the original and the demisyllable based synthesized speech without this rule. The result shows that this proposed rule can synthesize the more natural speech.

  • PDF