• Title/Summary/Keyword: Simulated speech

Search Result 70, Processing Time 0.027 seconds

FPGA Implementation of Speech Processor for Cochlear Implant (청각보철장치를 위한 어음 발췌기의 FPGA 구현)

  • Park, S.J.;Hong, M.S.;Shin, J.I.;Park, S.H.
    • Proceedings of the KOSOMBE Conference
    • /
    • v.1998 no.11
    • /
    • pp.163-164
    • /
    • 1998
  • In this paper the digital speech processing part of cochlear implant for sensorineural disorderly patients is implemented and simulated. We implement the speech processing part by dividing into three small parts - Filterbank, Pitch Detect, and Bandmapping parts. With the result, we conclude digital speech processing algorithm is implemented in FPGA perfectly. This means that cochlear implant can be made very small size.

  • PDF

Analysis on Vowel and Consonant Sounds of Patent's Speech with Velopharyngeal Insufficiency (VPI) and Simulated Speech (구개인두부전증 환자와 모의 음성의 모음과 자음 분석)

  • Sung, Mee Young;Kim, Heejin;Kwon, Tack-Kyun;Sung, Myung-Whun;Kim, Wooil
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.18 no.7
    • /
    • pp.1740-1748
    • /
    • 2014
  • This paper focuses on listening test and acoustic analysis of patients' speech with velopharyngeal insufficiency (VPI) and normal speakers' simulation speech. In this research, a set consisting of 50-words, vowels and single syllables is determined for speech database construction. A web-based listening evaluation system is developed for a convenient/automated evaluation procedure. The analysis results show the trend of incorrect recognition for VPI speech and the one for simulation speech are similar. Such similarity is also confirmed by comparing the formant locations of vowel and spectrum of consonant sounds. These results show that the simulation method for VPI speech is effective at generating the speech signals similar to actual VPI patient's speech. It is expected that the simulation speech data can be effectively employed for our future work such as acoustic model adaptation.

IMPLEMENTATION OF REAL TIME RELP VOCODER ON THE TMS320C25 DSP CHIP

  • Kwon, Kee-Hyeon;Chong, Jong-Wha
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • 1994.06a
    • /
    • pp.957-962
    • /
    • 1994
  • Real-time RELP vocoder is implemented on the TMS320C25 DSP chip. The implemented system is IBM-PC add-on board and composed of analog in/out unit, DSP unit, memoy unit, IBM-PC interface unit and its supporting assembly software. Speech analyzer and synthesizer is implimented by DSP assembly software. Speech parameters such as LPC coefficients, base-band residuals, and signal gains is extracted by autocorrelation method and inverse filter and synthesized by spectral folding method and direct form synthesis filter in this board. And then, real-time RELP vocoder with 9.6Kbps is simulated by down-loading method in the DSP program RAM.

  • PDF

Design of Emotion Recognition Model Using fuzzy Logic (퍼지 로직을 이용한 감정인식 모델설계)

  • 김이곤;배영철
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2000.05a
    • /
    • pp.268-282
    • /
    • 2000
  • Speech is one of the most efficient communication media and it includes several kinds of factors about speaker, context emotion and so on. Human emotion is expressed in the speech, the gesture, the physiological phenomena(the breath, the beating of the pulse, etc). In this paper, the method to have cognizance of emotion from anyone's voice signals is presented and simulated by using neuro-fuzzy model.

  • PDF

Design of Emotion Recognition Using Speech Signals (음성신호를 이용한 감정인식 모델설계)

  • 김이곤;김서영;하종필
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2001.10a
    • /
    • pp.265-270
    • /
    • 2001
  • Voice is one of the most efficient communication media and it includes several kinds of factors about speaker, context emotion and so on. Human emotion is expressed in the speech, the gesture, the physiological phenomena(the breath, the beating of the pulse, etc). In this paper, the method to have cognizance of emotion from anyone's voice signals is presented and simulated by using neuro-fuzzy model.

  • PDF

Analysis of Feature Extraction Methods for Distinguishing the Speech of Cleft Palate Patients (구개열 환자 발음 판별을 위한 특징 추출 방법 분석)

  • Kim, Sung Min;Kim, Wooil;Kwon, Tack-Kyun;Sung, Myung-Whun;Sung, Mee Young
    • Journal of KIISE
    • /
    • v.42 no.11
    • /
    • pp.1372-1379
    • /
    • 2015
  • This paper presents an analysis of feature extraction methods used for distinguishing the speech of patients with cleft palates and people with normal palates. This research is a basic study on the development of a software system for automatic recognition and restoration of speech disorders, in pursuit of improving the welfare of speech disabled persons. Monosyllable voice data for experiments were collected for three groups: normal speech, cleft palate speech, and simulated clef palate speech. The data consists of 14 basic Korean consonants, 5 complex consonants, and 7 vowels. Feature extractions are performed using three well-known methods: LPC, MFCC, and PLP. The pattern recognition process is executed using the acoustic model GMM. From our experiments, we concluded that the MFCC method is generally the most effective way to identify speech distortions. These results may contribute to the automatic detection and correction of the distorted speech of cleft palate patients, along with the development of an identification tool for levels of speech distortion.

The Influence of Feedback in the Simulated Patient Case-History Training among Audiology Students at the International Islamic University Malaysia

  • Dzulkarnain, Ahmad Aidil Arafat;Sani, Maryam Kamilah Ahmad;Rahmat, Sarah;Jusoh, Masnira
    • Journal of Audiology & Otology
    • /
    • v.23 no.3
    • /
    • pp.121-128
    • /
    • 2019
  • Background and Objectives: There is a scant evidence on the use of simulations in audiology (especially in Malaysia) for case-history taking, although this technique is widely used for training medical and nursing students. Feedback is one of the important components in simulations training; however, it is unknown if feedback by instructors could influence the simulated patient (SP) training outcome for case-history taking among audiology students. Aim of the present study is to determine whether the SP training with feedback in addition to the standard role-play and seminar training is an effective learning tool for audiology case-history taking. Subjects and Methods: Twenty-six second-year undergraduate audiology students participated. A cross-over study design was used. All students initially attended two hours of seminar and role-play sessions. They were then divided into three types of training, 1) SP training (Group A), 2) SP with feedback (Group B), and 3) a non-additional training group (Group C). After two training sessions, the students changed their types of training to, 1) Group A and C: SP training with feedback, and 2) Group B: non-additional training. All the groups were assessed at three points: 1) pre-test, 2) intermediate, and 3) post-test. The normalized median score differences between and within the respective groups were analysed using non-parametric tests at 95% confidence intervals. Results: Groups with additional SP trainings (with and without feedback) showed a significantly higher normalized gain score than no training group (p<0.05). Conclusions: The SP training (with/without feedback) is a beneficial learning tool for history taking to students in audiology major.

The Influence of Feedback in the Simulated Patient Case-History Training among Audiology Students at the International Islamic University Malaysia

  • Dzulkarnain, Ahmad Aidil Arafat;Sani, Maryam Kamilah Ahmad;Rahmat, Sarah;Jusoh, Masnira
    • Korean Journal of Audiology
    • /
    • v.23 no.3
    • /
    • pp.121-128
    • /
    • 2019
  • Background and Objectives: There is a scant evidence on the use of simulations in audiology (especially in Malaysia) for case-history taking, although this technique is widely used for training medical and nursing students. Feedback is one of the important components in simulations training; however, it is unknown if feedback by instructors could influence the simulated patient (SP) training outcome for case-history taking among audiology students. Aim of the present study is to determine whether the SP training with feedback in addition to the standard role-play and seminar training is an effective learning tool for audiology case-history taking. Subjects and Methods: Twenty-six second-year undergraduate audiology students participated. A cross-over study design was used. All students initially attended two hours of seminar and role-play sessions. They were then divided into three types of training, 1) SP training (Group A), 2) SP with feedback (Group B), and 3) a non-additional training group (Group C). After two training sessions, the students changed their types of training to, 1) Group A and C: SP training with feedback, and 2) Group B: non-additional training. All the groups were assessed at three points: 1) pre-test, 2) intermediate, and 3) post-test. The normalized median score differences between and within the respective groups were analysed using non-parametric tests at 95% confidence intervals. Results: Groups with additional SP trainings (with and without feedback) showed a significantly higher normalized gain score than no training group (p<0.05). Conclusions: The SP training (with/without feedback) is a beneficial learning tool for history taking to students in audiology major.

Speech Recognition with Image Information (영상정보 보완에 의한 음성인식)

  • 이천우;이상원;양근모;박인정
    • Proceedings of the IEEK Conference
    • /
    • 1999.06a
    • /
    • pp.511-515
    • /
    • 1999
  • The main factor decreasing speech recognition rate is the surrounding noise. To lower the noise effect, we generally used the filter bank at preprocessing stage. But, in this paper, we tried to recognize the 10 numeral numbers using 2-D LPC to extract image feature. At first, we obtained the result of speech-only recognition using 13th-order LPC coefficients and then, for distorted speech recognition results of ‘0’, ‘4’, ‘5’, ‘6’ and 9’, we added image parameters such as 12th-order 2-D LPC coefficients. At each frame, we extracted the 2-D LPC coefficients, and simulated recognizer with two parameters such as speech and image. Finally, for the numbers, such as ‘4’and ‘9’, the better results were obtained.

  • PDF

Sums-of-Products Models for Korean Segment Duration Prediction

  • Chung, Hyun-Song
    • Speech Sciences
    • /
    • v.10 no.4
    • /
    • pp.7-21
    • /
    • 2003
  • Sums-of-Products models were built for segment duration prediction of spoken Korean. An experiment for the modelling was carried out to apply the results to Korean text-to-speech synthesis systems. 670 read sentences were analyzed. trained and tested for the construction of the duration models. Traditional sequential rule systems were extended to simple additive, multiplicative and additive-multiplicative models based on Sums-of-Products modelling. The parameters used in the modelling include the properties of the target segment and its neighbors and the target segment's position in the prosodic structure. Two optimisation strategies were used: the downhill simplex method and the simulated annealing method. The performance of the models was measured by the correlation coefficient and the root mean squared prediction error (RMSE) between actual and predicted duration in the test data. The best performance was obtained when the data was trained and tested by ' additive-multiplicative models. ' The correlation for the vowel duration prediction was 0.69 and the RMSE. 31.80 ms. while the correlation for the consonant duration prediction was 0.54 and the RMSE. 29.02 ms. The results were not good enough to be applied to the real-time text-to-speech systems. Further investigation of feature interactions is required for the better performance of the Sums-of-Products models.

  • PDF