• Title/Summary/Keyword: Acoustic Features

Search Result 328, Processing Time 0.026 seconds

Broadband Noise Prediction of the Ice-maker Centrifugal Fan in a Refrigerator Using Hybrid CAA Method and FRPM Technique (복합 CAA 방법과 FRPM 기법을 이용한 냉장고 얼음제조용 원심팬의 광대역 소음 예측)

  • Heo, Seung;Kim, Dae-Hwan;Cheong, Cheol-Ung
    • The Journal of the Acoustical Society of Korea
    • /
    • v.31 no.6
    • /
    • pp.391-398
    • /
    • 2012
  • In this paper, prediction of centrifugal fan was conducted through combination the hybrid CAA method which was used to predict the fan noise with the FRPM technique which was used to generate the broadband noise source. Firstly, flow field surround the centrifugal fan was computed using the RANS equations and noise source region was deducted from the computed flow field. Then the FRPM technique was applied to the source region for generation of turbulence which satisfies the stochastic features. The noise source of the centrifugal fan was modeled by applying the acoustic analogy to the synthesized flow field from the computed and generated flow fields. Finally, the broadband noise of the centrifugal fan was predicted through combination the modeled noise source with the linear propagation which was realized using the boundary element method. It was confirmed that the proposed technique is efficient to predict the tonal and broadband noises of centrifugal fan through comparison with the measured data.

Room Acoustic Design in International Convention Center Jeju (제주국제컨벤션센터 컨퍼런스홀의 건축음향 설계)

  • 주현경;오양기;두세진;김하근
    • Proceedings of the Korean Society for Noise and Vibration Engineering Conference
    • /
    • 2003.05a
    • /
    • pp.802-807
    • /
    • 2003
  • International Convention Center Jeju(ICCJ) was planed and built for accommodating a variety of conventional and exhibitional activities. For a better flexibility of operation, almost all rooms in ICCJ are designed to be subdivided Into a couple of small rooms with installation of movable partition walls. Architectural and acoustical design should be deliberatively and cooperatively undertaken to cope with such a complex condition. Conference hall, the largest room in ICCJ, has a capacity of 5000 seats who:1 used as a whole. It could be divided into 3 halls, one with 2000 pre-installed seats on slanted floor, up/down removable stage and its settings above, and the other 2 flat rooms with no seats installed. A devided hall with pre-installed seats and stage is designed for a multi-use auditorium. Almost all surfaces except ceilings adjacent to the stage are sound absorptively treated, in regard to extensive use of sound reinforcement systems. Its reverberation time 1.65 sec without audience, which is roughly correspond to 1.50 sec with fully occupied audience. When there is a need for a larger room, all the partition wail Is removed and the hall could be used as a whole. Exhibition hall is located in the first floor of ICCJ. Absorption and softness are needed for the hat 1 because exhibition behavior has something noisy features. Perforated MDF panels with porous materials and air space in the back groundare adopted for the walls. There are one large, two medium, and several small convention rooms in ICJJ. The room are also acoustically designed for maximum flexibility with no defects soundwisely.

  • PDF

The Comparison of the Acoustic and Aerodynamic Characteristics of $PROVOX^{(R)}$ Voice and Esophageal Voice Produced by the Same Laryngectomee (동일 후적자가 산출하는 기관식도 발성($PROVOX^{(R)}$ 발성)과 식도 발성에 대한 음향학적 및 공기역학적 특성 비교)

  • Pyo, H.Y.;Choi, H.S.;Lim, S.E.;Choi, S.H.
    • Speech Sciences
    • /
    • v.5 no.1
    • /
    • pp.121-139
    • /
    • 1999
  • Our experimental subject was a laryngectomee who had undergone total laryngectomy with $PROVOX^{(R)}$ insertion, and learned esophageal speech after the surgery, so he could produce both $PROVOX^{(R)}$ voice and esophageal voice. With this subject's production of $PROVOX^{(R)}$ and esophageal voice, we are to compare the acoustic and aerodynamic characteristics of the two voices, under the same physical conditions of the same person. As a result, the fundamental frequency of esophageal voice was 137.2 Hz, and that of $PROVOX^{(R)}$ was 97.5 Hz. $PROVOX^{(R)}$ voice showed lower jitter, shimmer and NHR than esophageal voice, which means that $PROVOX^{(R)}$ voice showed better voice quality than esophageal voice. In spectrographic analysis, the formation of formants and pseudoformants were more distinct in esophageal voice and several temporal aspects of acoutic features such as VOT and closure duration were more similar with normal voice in $PROVOX^{(R)}$ voice. During the sentence utterance, esophageal voice showed longer pause or silence duration than $PROVOX^{(R)}$ voice. Maximum phonation time and mean flow rate of $PROVOX^{(R)}$ voice were much longer and larger than esophageal voice, but mean and range of sound pressure level, subglottic pressure and voice efficiency were similar in the two voices. Glottal resistance of esophageal voice was much larger than $PROVOX^{(R)}$ voice which showed still larger glottal resistance than normal voice.

  • PDF

Speech synthesis using acoustic Doppler signal (초음파 도플러 신호를 이용한 음성 합성)

  • Lee, Ki-Seung
    • The Journal of the Acoustical Society of Korea
    • /
    • v.35 no.2
    • /
    • pp.134-142
    • /
    • 2016
  • In this paper, a method synthesizing speech signal using the 40 kHz ultrasonic signals reflected from the articulatory muscles was introduced and performance was evaluated. When the ultrasound signals are radiated to articulating face, the Doppler effects caused by movements of lips, jaw, and chin observed. The signals that have different frequencies from that of the transmitted signals are found in the received signals. These ADS (Acoustic-Doppler Signals) were used for estimating of the speech parameters in this study. Prior to synthesizing speech signal, a quantitative correlation analysis between ADS and speech signals was carried out on each frequency bin. According to the results, the feasibility of the ADS-based speech synthesis was validated. ADS-to-speech transformation was achieved by the joint Gaussian mixture model-based conversion rules. The experimental results from the 5 subjects showed that filter bank energy and LPC (Linear Predictive Coefficient) cepstrum coefficients are the optimal features for ADS, and speech, respectively. In the subjective evaluation where synthesized speech signals were obtained using the excitation sources extracted from original speech signals, it was confirmed that the ADS-to-speech conversion method yielded 72.2 % average recognition rates.

Efficacy of laughing voice treatment (SKMVTT) in benign vocal fold lesions (양성성대질환의 웃음 음성치료(SKMVTT))

  • Jung, Dae-Yong;Wi, Joon-Yeol;Kim, Seong-Tae
    • Phonetics and Speech Sciences
    • /
    • v.10 no.4
    • /
    • pp.155-161
    • /
    • 2018
  • The purpose of this study was to evaluate the efficacy of a multiple voice therapy technique ($SKMVTT^{(R)}$) using laughter for the treatment of various benign vocal fold lesions. To achieve this, 23 female patients diagnosed with vocal nodules, vocal polyp, and muscle tension dysphonia through videostroboscopy were enrolled in vocal hygiene and $SKMVTT^{(R)}$. All of the patients were treated once a week for 4 to 12 sessions. The GRBAS scale was used to confirm the changes in voice quality before and after the treatment. Acoustic analysis was performed to evaluate jitter, shimmer, NHR, fundamental frequency variation, amplitude variation, PFR, and dB range. Videostroboscopy was performed to confirm the changes in the laryngeal features before and after the treatment. After the $SKMVTT^{(R)}$, the results of the perceptual evaluation demonstrated that the G, R, and B scales significantly improved. An acoustic evaluation also demonstrated that jitter, shimmer, NHR, vAm, vFo, PFR, and dB range also significantly improved after the $SKMVTT^{(R)}$. In comparison to the videostroboscopic findings, the size of the vocal nodules and vocal polyp decreased or disappeared after the treatment. In addition, the size of the cuneiform tubercles decreased, the length of the aryepiglottic folds became longer, and the laryngeal findings of the supraglottic compressions improved after the $SKMVTT^{(R)}$. These results suggest that the $SKMVTT^{(R)}$ is effective in improving the vocal quality of patients with benign vocal fold lesions. In conclusion, it seems that laughter and inspiratory phonation suppressed abnormal laryngeal elevation and lowered laryngeal height, which seems to have the effect of improving hyperfunctional phonation.

Investigation of 0.5 MJ superconducting energy storage system by acoustic emission method.

  • Miklyaev, S.M.;Shevchenko, S.A.;Surin, M.I.
    • Proceedings of the KIPE Conference
    • /
    • 1998.10a
    • /
    • pp.961-965
    • /
    • 1998
  • The rapid development of small-scale (1-10 MJ) Superconducting Magnetic Energy Storage Systems (SMES) can be explained by real perspective of practical implementation of these devices in electro power nets. However the serious problem of all high mechanically stressed superconducting coils-problem of training and degradation (decreasing) of operating current still exists. Moreover for SMES systems this problems is more dangerous because of pulsed origin of mechanical stresses-one of the major sources of local heat disturbances in superconducting coils. We investigated acoustic emission (AE) phenomenon on model and 0.5 MJ SMES coils taking into account close correlation of AE and local heat disturbances. Two-coils 0.5 MJ SMES system was developed, manufactured and tested at Russian Research Center in the frames of cooperation with Korean Electrical Engineering Company (KEPCO) [1]. The two-coil SMES operates with the stored energy transmitted between coils in the course of a single cycle with 2 seconds energy transfer time. Maximum operating current 1.55 kA corresponds to 0.5 MF in each coil. The Nb-Ti-based conductor was designed and used for SMES manufacturing. It represents transposed cable made of Nb-Ti strands in copper matrix, several cooper strands and several stainless steel strands. The coils are wound onto fiberglass cylindrical bobbins. To make AE event information more useful a real time instrumentation system was used. Two main measured and computer processed AE parameters were considered: the energy of AE events (E) and the accumulated energy of AE events (E ). Influence of current value in 0.5 MJ coils on E and E was studied. The sensors were installed onto the bobbin and the external surface of magnets. Three levels of initial current were examined: 600A, 1000A, 2450 A. An extraordinary strong dependence of the current level on E and E was observed. The specific features of AE from model coils, operated in sinusoidal vibration current changing mode were investigated. Three current frequency modes were examined: 0.012 Hz, 0.03 Hz and 0.12 Hz. In all modes maximum amplitude 1200 A was realized.

  • PDF

Aerodynamic Features and Voice Therapy Interventions of Functional Voice Disorder after Thyroidectomy (갑상선 절제 술 후 기능적 음성장애의 공기역학적 특징과 음성치료 중재)

  • Lee, Chang-Yoon;An, Soo-Youn;Chang, Hyun;Jeong, Hee Seok;Son, Hee Young
    • Journal of the Korean Society of Laryngology, Phoniatrics and Logopedics
    • /
    • v.26 no.1
    • /
    • pp.25-33
    • /
    • 2015
  • Background and Objectives:The objective of this study was to investigate the features of post-thyroidectomy subjective voice disorder by Voice Handicap Index (VHI) and Voice Symptom Scale (VOISS) through aerodynamic analysis and to investigate the appropriate voice therapy intervention. Materials and Methods:Twenty post-thyroidectomy patients who had no recurrent laryngeal nerve paralysis through laryngeal stroboscopy were enrolled for this study. Acoustic and aerodynamic evaluations were performed before operation, 2 weeks and 3 months after operation. Subjective voice evaluation was performed by VHI and VOISS. Aerodynamic evaluation was compared and analysed by maximum phonation time(MPT), phonation threshold pressure(PTP), mean air flow rate(MFR), etc. Subjective voice evaluation was surveyed through VHI and VOISS. To evaluate patients' symptoms related to functional voice disorder, scores on physical domain in VHI and VOISS were selected to be compared for each session. Results: The 10 out of 20 participants who complained of voice symptoms had no significant difference with pre-operation in acoustic evaluation, but all showed higher scores on 2 weeks and 3 months after operation compared to pre-operation, in VHI-physical domain and selected questionnaires in VOISS. They reduced MPT and increased PTP value simultaneously. Laryngeal massage and breathing training were simultaneously treated to them, 5 participants resulting in improvement in MPT and PTP compared to pre-treatment. Conclusion:Patients who complained voice change with no organic damage after thyroidectomy were all shown to have reduced MPT and increased PTP in some by aerodynamic evaluations. Reduced MPT may imply some problem in air flow beneath glottis. Increased PTP suggests much more effort in vocalization mechanism than pre-operation. Comparing aerodynamic evaluations in post-thyroidectomy may provide information on behavioral interventions. Additionally, study on laryngeal massage and breathing training simultaneously treated to patients with such voice disorder is needed to be conducted with larger number of participants.

  • PDF

A study of the prosodic patterns of autism and normal children in the imitating declarative and interrogative sentences (따라말하기 과제를 통한 자폐범주성 장애 아동과 일반 아동의 평서문과 의문문의 음향학적 특성 비교)

  • Lee, Jinhyung;Seong, Cheoljae
    • Phonetics and Speech Sciences
    • /
    • v.12 no.2
    • /
    • pp.39-49
    • /
    • 2020
  • The prosody of children with autism spectrum disorders (ASD) has several abnormal features, including monotonous speech. The purpose of this study was to compare acoustic features between an ASD group and a typically developing (TD) group and within the ASD group. The study also examined audience perceptions of the lengthening effect of increasing the number of syllables. 50 participants were divided into two groups (20 with ASD and 30 TD), and they were asked to imitate a total of 28 sentences. In the auditory-perceptual evaluation, seven participants chose sentence types in 115 sentences. Pitch, intensity, speech rate, and pitch slope were used to analyze the significant differences. In conclusion, the ASD group showed higher pitch and intensity and a lower overall speaking rate than the TD group. Moreover, there were significant differences in s2 slope of interrogative sentences. Finally, based on the auditory-perceptual evaluation, only 4.3% of interrogative sentences produced by participants with ASD were perceived as declarative sentences. The cause of this abnormal prosody has not been clearly identified; however, pragmatic ability and other characteristics of autism are related to ASD prosody. This study identified prosodic ASD patterns and suggested the need to develop treatments to improve prosody.

Speech Visualization of Korean Vowels Based on the Distances Among Acoustic Features (음성특징의 거리 개념에 기반한 한국어 모음 음성의 시각화)

  • Pok, Gouchol
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.12 no.5
    • /
    • pp.512-520
    • /
    • 2019
  • It is quite useful to represent speeches visually for learners who study foreign languages as well as the hearing impaired who cannot directly hear speeches, and a number of researches have been presented in the literature. They remain, however, at the level of representing the characteristics of speeches using colors or showing the changing shape of lips and mouth using the animation-based representation. As a result of such approaches, those methods cannot tell the users how far their pronunciations are away from the standard ones, and moreover they make it technically difficult to develop such a system in which users can correct their pronunciation in an interactive manner. In order to address these kind of drawbacks, this paper proposes a speech visualization model based on the relative distance between the user's speech and the standard one, furthermore suggests actual implementation directions by applying the proposed model to the visualization of Korean vowels. The method extract three formants F1, F2, and F3 from speech signals and feed them into the Kohonen's SOM to map the results into 2-D screen and represent each speech as a pint on the screen. We have presented a real system implemented using the open source formant analysis software on the speech of a Korean instructor and several foreign students studying Korean language, in which the user interface was built using the Javascript for the screen display.

Hi, KIA! Classifying Emotional States from Wake-up Words Using Machine Learning (Hi, KIA! 기계 학습을 이용한 기동어 기반 감성 분류)

  • Kim, Taesu;Kim, Yeongwoo;Kim, Keunhyeong;Kim, Chul Min;Jun, Hyung Seok;Suk, Hyeon-Jeong
    • Science of Emotion and Sensibility
    • /
    • v.24 no.1
    • /
    • pp.91-104
    • /
    • 2021
  • This study explored users' emotional states identified from the wake-up words -"Hi, KIA!"- using a machine learning algorithm considering the user interface of passenger cars' voice. We targeted four emotional states, namely, excited, angry, desperate, and neutral, and created a total of 12 emotional scenarios in the context of car driving. Nine college students participated and recorded sentences as guided in the visualized scenario. The wake-up words were extracted from whole sentences, resulting in two data sets. We used the soundgen package and svmRadial method of caret package in open source-based R code to collect acoustic features of the recorded voices and performed machine learning-based analysis to determine the predictability of the modeled algorithm. We compared the accuracy of wake-up words (60.19%: 22%~81%) with that of whole sentences (41.51%) for all nine participants in relation to the four emotional categories. Accuracy and sensitivity performance of individual differences were noticeable, while the selected features were relatively constant. This study provides empirical evidence regarding the potential application of the wake-up words in the practice of emotion-driven user experience in communication between users and the artificial intelligence system.