• Title/Summary/Keyword: speech sound

Search Result 625, Processing Time 0.023 seconds

Acoustic features of diphthongs produced by children with speech sound disorders (말소리장애 아동이 산출한 이중모음의 음향학적 특성)

  • Cho, Yoon Soo;Pyo, Hwa Young;Han, Jin Soon;Lee, Eun Ju
    • Phonetics and Speech Sciences
    • /
    • v.13 no.1
    • /
    • pp.65-72
    • /
    • 2021
  • The aim of this study is to prepare basic data that can be used for evaluation and intervention by investigating the characteristics of diphthongs produced by children with speech sound disorders. To confirm this, two groups of 10 children each, with and without speech sound disorders were asked to imitate the meaningless two-syllable 'diphthongs + da'. The slope of F1 and F2, amount of change of formant, and duration of glide were analyzed by Praat (version 6.1.16). As a result, the difference between the two groups was found in the slope of F1 of /ju/. Children with speech sound disorders had smaller changes in formants and shorter duration time values compared to normal children, and there were statistically significant differences. The amount of change in formant in the glide was found in F1 of /ju, jɛ/, F2 of /jɑ, jɛ/, and there were significant differences in the duration of glide in /ju, jɛ/. The results of this study showed that the range of articulation of diphthongs in children with speech sound disorders is relatively smaller than that of normal children, thus the time it takes to articulate was reduced. These results suggest that the range of articulation and acoustic analysis should be further investigated for evaluation and intervention regarding diphthongs of children with speech sound disorders.

Sound-Field Speech Evoked Auditory Brainstem Response in Cochlear-Implant Recipients

  • Jarollahi, Farnoush;Valadbeigi, Ayub;Jalaei, Bahram;Maarefvand, Mohammad;Zarandy, Masoud Motasaddi;Haghani, Hamid;Shirzhiyan, Zahra
    • Korean Journal of Audiology
    • /
    • v.24 no.2
    • /
    • pp.71-78
    • /
    • 2020
  • Background and Objectives: Currently limited information is available on speech stimuli processing at the subcortical level in the recipients of cochlear implant (CI). Speech processing in the brainstem level is measured using speech-auditory brainstem response (S-ABR). The purpose of the present study was to measure the S-ABR components in the sound-field presentation in CI recipients, and compare with normal hearing (NH) children. Subjects and Methods: In this descriptive-analytical study, participants were divided in two groups: patients with CIs; and NH group. The CI group consisted of 20 prelingual hearing impairment children (mean age=8.90±0.79 years), with ipsilateral CIs (right side). The control group consisted of 20 healthy NH children, with comparable age and sex distribution. The S-ABR was evoked by the 40-ms synthesized /da/ syllable stimulus that was indicated in the sound-field presentation. Results: Sound-field S-ABR measured in the CI recipients indicated statistically significant delayed latencies, than in the NH group. In addition, these results demonstrated that the frequency following response peak amplitude was significantly higher in CI recipients, than in the NH counterparts (p<0.05). Finally, the neural phase locking were significantly lower in CI recipients (p<0.05). Conclusions: The findings of sound-field S-ABR demonstrated that CI recipients have neural encoding deficits in temporal and spectral domains at the brainstem level; therefore, the sound-field S-ABR can be considered an efficient clinical procedure to assess the speech process in CI recipients.

Sound-Field Speech Evoked Auditory Brainstem Response in Cochlear-Implant Recipients

  • Jarollahi, Farnoush;Valadbeigi, Ayub;Jalaei, Bahram;Maarefvand, Mohammad;Zarandy, Masoud Motasaddi;Haghani, Hamid;Shirzhiyan, Zahra
    • Journal of Audiology & Otology
    • /
    • v.24 no.2
    • /
    • pp.71-78
    • /
    • 2020
  • Background and Objectives: Currently limited information is available on speech stimuli processing at the subcortical level in the recipients of cochlear implant (CI). Speech processing in the brainstem level is measured using speech-auditory brainstem response (S-ABR). The purpose of the present study was to measure the S-ABR components in the sound-field presentation in CI recipients, and compare with normal hearing (NH) children. Subjects and Methods: In this descriptive-analytical study, participants were divided in two groups: patients with CIs; and NH group. The CI group consisted of 20 prelingual hearing impairment children (mean age=8.90±0.79 years), with ipsilateral CIs (right side). The control group consisted of 20 healthy NH children, with comparable age and sex distribution. The S-ABR was evoked by the 40-ms synthesized /da/ syllable stimulus that was indicated in the sound-field presentation. Results: Sound-field S-ABR measured in the CI recipients indicated statistically significant delayed latencies, than in the NH group. In addition, these results demonstrated that the frequency following response peak amplitude was significantly higher in CI recipients, than in the NH counterparts (p<0.05). Finally, the neural phase locking were significantly lower in CI recipients (p<0.05). Conclusions: The findings of sound-field S-ABR demonstrated that CI recipients have neural encoding deficits in temporal and spectral domains at the brainstem level; therefore, the sound-field S-ABR can be considered an efficient clinical procedure to assess the speech process in CI recipients.

Performance Estimation of a Window Shaker (유리창 도청방지 장치의 성능평가)

  • Kim, Seock-Hyun;Kim, Hee-Dong;Heo, Wook
    • Proceedings of the Korean Society for Noise and Vibration Engineering Conference
    • /
    • 2007.05a
    • /
    • pp.649-654
    • /
    • 2007
  • Eavesdropping prevention performance is evaluated on a commercial window shaker, which is used to prevent a glass window from eavesdropping. Speech transmission index (STI) is introduced in order to estimate quantitatively the speech intelligibility of the sound detected on the glass window. Objective test by IEC standard using modulation transfer function (MTF) is performed to determine STI. Using Maximum Length Sequency (MLS) signal as a sound source, MTF is measured by accelerometers and laser doppler vibrometer. STI under different level of disturbing wave are compared to confirm the disturbing effect on the speech intelligibility.

  • PDF

The Emotion Recognition System through The Extraction of Emotional Components from Speech (음성의 감성요소 추출을 통한 감성 인식 시스템)

  • Park Chang-Hyun;Sim Kwee-Bo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.10 no.9
    • /
    • pp.763-770
    • /
    • 2004
  • The important issue of emotion recognition from speech is a feature extracting and pattern classification. Features should involve essential information for classifying the emotions. Feature selection is needed to decompose the components of speech and analyze the relation between features and emotions. Specially, a pitch of speech components includes much information for emotion. Accordingly, this paper searches the relation of emotion to features such as the sound loudness, pitch, etc. and classifies the emotions by using the statistic of the collecting data. This paper deals with the method of recognizing emotion from the sound. The most important emotional component of sound is a tone. Also, the inference ability of a brain takes part in the emotion recognition. This paper finds empirically the emotional components from the speech and experiment on the emotion recognition. This paper also proposes the recognition method using these emotional components and the transition probability.

A new sound source localization method robust to microphones' gain (마이크로폰의 이득 특성에 강인한 위치 추적)

  • Choi Ji-Sung;Lee Ji-Yeoun;Jeong Sang-Bae;Hahn Min-Soo
    • Proceedings of the KSPS conference
    • /
    • 2006.05a
    • /
    • pp.127-130
    • /
    • 2006
  • This paper suggests an algorithm that can estimate the direction of the sound source with three microphones arranged on a circle. The algorithm is robust to microphones' gains because it uses only the time differences between microphones. To make this possible, a cost function which normalizes the microphone's gains is utilized and a procedure to detect the rough position of the sound source is also proposed. Through our experiments, we obtained significant performance improvement compared with the energy-based localizer.

  • PDF

Multi-Pulse Amplitude and Location Estimation by Maximum-Likelihood Estimation in MPE-LPC Speech Synthesis (MPE-LPC음성합성에서 Maximum- Likelihood Estimation에 의한 Multi-Pulse의 크기와 위치 추정)

  • 이기용;최홍섭;안수길
    • Journal of the Korean Institute of Telematics and Electronics
    • /
    • v.26 no.9
    • /
    • pp.1436-1443
    • /
    • 1989
  • In this paper, we propose a maximum-likelihood estimation(MLE) method to obtain the location and the amplitude of the pulses in MPE( multi-pulse excitation)-LPC speech synthesis using multi-pulses as excitation source. This MLE method computes the value maximizing the likelihood function with respect to unknown parameters(amplitude and position of the pulses) for the observed data sequence. Thus in the case of overlapped pulses, the method is equivalent to Ozawa's crosscorrelation method, resulting in equal amount of computation and sound quality with the cross-correlation method. We show by computer simulation: the multi-pulses obtained by MLE method are(1) pseudo-periodic in pitch in the case of voicde sound, (2) the pulses are random for unvoiced sound, (3) the pulses change from random to periodic in the interval where the original speech signal changes from unvoiced to voiced. Short time power specta of original speech and syunthesized speech obtained by using multi-pulses as excitation source are quite similar to each other at the formants.

  • PDF

VOICE CONTROL SYSTEM FOR TELEVISION SET USING MASKING MODEL AS A FRONT-END OF SPEECH RECOGNIZER

  • Usagawa, Tsuyoshi;Iwata, Makoto;Ebata, Masanao
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • 1994.06a
    • /
    • pp.991-996
    • /
    • 1994
  • Surrounding noise often affects the performance of speech recognition system when it is used in office or home. Especially situation is more serious when colored and nonstational noise such as an sound from television or other audio equipment is introduced. The authors proposed a voice control system for television set using an adaptive noise canceler, and it works well even is sound of television set has comparable level of speech. In this paper, a new front-end of speech recognition is introduced for the voice control system. This font-end utilizes a simplified masking model to reduce the effect of residual noise. According to experimental results, 90% correct recognition is achieved even if the level of television sound is almost 15dB higher than one of speech.

  • PDF

A Emergency Sound Detecting Method for Smarter City (스마트 시티에서의 이머전시 사운드 감지방법)

  • Cho, Young-Im
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.16 no.12
    • /
    • pp.1143-1149
    • /
    • 2010
  • Because the noise is the main cause for decreasing the performance at speech recognition, the place or environment is very important in speech recognition. To improve the speech recognition performance in the real situations where various extraneous noises are abundant, a novel combination of FIR and Wiener filters is proposed and experimented. The combination resulted in improved accuracy and reduced processing time, enabling fast analysis and response in emergency situations. Usually, there are many dangerous situations in our city life, so for the smarter city it is necessary to detect many types of sound in various environment. Therefore this paper is about how to detect many types of sound in real city, especially on CCTV. This paper is for implementing the smarter city by detecting many types of sounds and filtering one of the emergency sound in this sound stream. And then it can be possible to handle with the emergency or dangerous situation.

Sound Source Localization using Acoustically Shadowed Microphones (가려진 마이크로폰을 이용한 음원 위치 추적)

  • Lee, Hyeop-Woo;Yook, Dong-Suk
    • Speech Sciences
    • /
    • v.15 no.3
    • /
    • pp.17-28
    • /
    • 2008
  • In many practical applications of robots, finding the location of an incoming sound is an important issue for the development of efficient human robot interface. Most sound source localization algorithms make use of only those microphones that are acoustically visible from the sound source or do not take into account the effect of sound diffraction, thereby degrading the sound source localization performance. This paper proposes a new sound source localization method that can utilize those microphones that are acoustically shadowed from the sound source. The experiment results show that use of the acoustically shadowed microphones, which receive higher signal-to-noise ratio signals than the others and are closer to the sound source, improves the performance of sound source localization.

  • PDF