• Title/Summary/Keyword: Speech Understanding

Search Result 188, Processing Time 0.025 seconds

A comparative study between French schwa and Korean [i] - An experimental phonetic and phonological perspective -

  • Lee, Eun-Yung;Kim, Seon-Jung
    • Speech Sciences
    • /
    • v.7 no.1
    • /
    • pp.171-186
    • /
    • 2000
  • The aim of this paper is to investigate the acoustic characteristics of the French vowel [e] and Korean [i] and to seek a way of understanding them from a phonological point of view. These two vowels have similar distributional properties, i.e. they alternate with zero in some contexts. Therefore, in both languages, they are not found when immediately followed by a nucleus with phonetic content and in word-final positions. We firstly compare the two vowels by measuring the actual frequencies of the formants, pitch and energy using CSL. We also consider whether the realisation of the two vowels is affected by the speed of speech sounds. In order to show that realisation of the two vowels in both languages is not arbitrary, rather predicted, we will introduce the notion of proper government, proposed and developed by Kaye (1987, 1990) and Charette (1991).

  • PDF

The Influence of Non-Linear Frequency Compression on the Perception of Speech and Music in Patients with High Frequency Hearing Loss

  • Ahn, Jungmin;Choi, Ji Eun;Kang, Ju Yong;Choi, Ik Joon;Lee, Myung-Chul;Lee, Byeong-Cheol;Hong, Sung Hwa;Moon, Il Joon
    • Korean Journal of Audiology
    • /
    • v.25 no.2
    • /
    • pp.80-88
    • /
    • 2021
  • Background and Objectives: Non-linear frequency compression (NLFC) technology compresses and shifts higher frequencies into a lower frequency area that has better residual hearing. Because consonants are uttered in the high-frequency area, NLFC could provide better speech understanding. The aim of this study was to investigate the clinical effectiveness of NLFC technology on the perception of speech and music in patients with high-frequency hearing loss. Subjects and Methods: Twelve participants with high-frequency hearing loss were tested in a counter-balanced order, and had two weeks of daily experience with NLFC set on/off prior to testing. Performance was repeatedly evaluated with consonant tests in quiet and noise environments, speech perception in noise, music perception and acceptableness of sound quality rating tasks. Additionally, two questionnaires (the Abbreviated Profile of Hearing Aid Benefit and the Korean version of the International Outcome Inventory-Hearing Aids) were administered. Results: Consonant and speech perception improved with hearing aids (NLFC on/off conditions), but there was no significant difference between NLFC on and off states. Music perception performances revealed no notable difference among unaided and NLFC on and off states. The benefits and satisfaction ratings between NLFC on and off conditions were also not significantly different, based on questionnaires, however great individual variability preferences were noted. Conclusions: Speech perception as well as music perception both in quiet and noise environments was similar between NLFC on and off states, indicating that real world benefits from NLFC technology may be limited in Korean adult hearing aid users.

The Influence of Non-Linear Frequency Compression on the Perception of Speech and Music in Patients with High Frequency Hearing Loss

  • Ahn, Jungmin;Choi, Ji Eun;Kang, Ju Yong;Choi, Ik Joon;Lee, Myung-Chul;Lee, Byeong-Cheol;Hong, Sung Hwa;Moon, Il Joon
    • Journal of Audiology & Otology
    • /
    • v.25 no.2
    • /
    • pp.80-88
    • /
    • 2021
  • Background and Objectives: Non-linear frequency compression (NLFC) technology compresses and shifts higher frequencies into a lower frequency area that has better residual hearing. Because consonants are uttered in the high-frequency area, NLFC could provide better speech understanding. The aim of this study was to investigate the clinical effectiveness of NLFC technology on the perception of speech and music in patients with high-frequency hearing loss. Subjects and Methods: Twelve participants with high-frequency hearing loss were tested in a counter-balanced order, and had two weeks of daily experience with NLFC set on/off prior to testing. Performance was repeatedly evaluated with consonant tests in quiet and noise environments, speech perception in noise, music perception and acceptableness of sound quality rating tasks. Additionally, two questionnaires (the Abbreviated Profile of Hearing Aid Benefit and the Korean version of the International Outcome Inventory-Hearing Aids) were administered. Results: Consonant and speech perception improved with hearing aids (NLFC on/off conditions), but there was no significant difference between NLFC on and off states. Music perception performances revealed no notable difference among unaided and NLFC on and off states. The benefits and satisfaction ratings between NLFC on and off conditions were also not significantly different, based on questionnaires, however great individual variability preferences were noted. Conclusions: Speech perception as well as music perception both in quiet and noise environments was similar between NLFC on and off states, indicating that real world benefits from NLFC technology may be limited in Korean adult hearing aid users.

Analogical Reasoning Skills and Metaphoric Understanding in School-age Children with Language Disabilities (학령기 언어장애아동의 유추추론능력과 은유이해능력 간의 상관관계연구)

  • Shin, Hu-Nam;Kwon, Do-Ha
    • MALSORI
    • /
    • no.67
    • /
    • pp.17-31
    • /
    • 2008
  • This study was designed to explore the analogical reasoning and metaphoric understanding in typically developing children and language impaired children. 13 Language-impaired children were matched to 16 typically developing children on the basis of receptive vocabulary age. All 29 children were enrolled in the 1st to 3rd grade in regular elementary schools. All were administered analogical reasoning and metaphoric tasks. Results indicated that the children with language disabilities did not perform as well as the receptive vocabulary matched group on the two tasks. In addition, we found that both of children with and without language disabilities did not have relationship between analogical reasoning and metaphoric understanding.

  • PDF

A Corpus-based study on the Effects of Gender on Voiceless Fricatives in American English

  • Yoon, Tae-Jin
    • Phonetics and Speech Sciences
    • /
    • v.7 no.1
    • /
    • pp.117-124
    • /
    • 2015
  • This paper investigates the acoustic characteristics of English fricatives in the TIMIT corpus, with a special focus on the role of gender in rendering fricatives in American English. The TIMIT database includes 630 talkers and 2342 different sentences, comprising over five hours of speech. Acoustic analyses are conducted in the domain of spectral and temporal properties by treating gender as an independent factor. The results of acoustic analyses revealed that the most acoustic properties of voiceless sibilants turned out to be different between male and female speakers, but those of voiceless non-sibilants did not show differences. A classification experiment using linear discriminant analysis (LDA) revealed that 85.73% of voiceless fricatives are correctly classified. The sibilants are 88.61% correctly classified, whereas the non-sibilants are only 57.91% correctly classified. The majority of the errors are from the misclassification of /ɵ/ as [f]. The average accuracy of gender classification is 77.67%. Most of the inaccuracy results are from the classification of female speakers in non-sibilants. The results are accounted for by resorting to biological differences as well as macro-social factors. The paper contributes to the understanding of the role of gender in a large-scale speech corpus.

Formant frequency changes of female voice /a/, /i/, /u/ in real ear (실이에서 여자 음성 /ㅏ/, /ㅣ/, /ㅜ/의 포먼트 주파수 변화)

  • Heo, Seungdeok;Kang, Huira
    • Phonetics and Speech Sciences
    • /
    • v.9 no.1
    • /
    • pp.49-53
    • /
    • 2017
  • Formant frequencies depend on the position of tongue, the shape of lips, and larynx. In the auditory system, the external ear canal is an open-end resonator, which can modify the voice characteristics. This study investigates the effect of the real ear on formant frequencies. Fifteen subjects ranging from 22 to 30 years of age participated in the study. This study employed three corner vowels: the low central vowel /a/, the high front vowel /i/, and the high back vowel /u/. For this study, the voice of a well-educated undergraduate who majored in speech-language pathology, was recorded with a high performance condenser microphone placed in the upper pinna and in the ear canal. Paired t-test showed that there were significant difference in the formant frequencies of F1, F2, F3, and F4 between the free field and the real ear. For /a/, all formant frequencies decreased significantly in the real ear. For /i/, F2 increased and F3 and F4 decreased. For /u/, F1 and F2 increased, but F3 and F4 decreased. It seems that these voice modifications in the real ear contribute to interpreting voice quality and understanding speech, timbre, and individual characteristics, which are influenced by the shape of the outer ear and external ear canal in such a way that formant frequencies become centralized in the vowel space.

The Diagnosis and Management of Velopharyngeal Insufficiency (연구개인두 폐쇄 부전 환자의 진단과 치료)

  • Lee, Yong-Kwon;Choi, Jae-Pyong;Choi, Jin-Young
    • Korean Journal of Cleft Lip And Palate
    • /
    • v.11 no.1
    • /
    • pp.13-22
    • /
    • 2008
  • Velopharyngeal insufficiency(VPI), characterized by hypernasal resonance and nasal air emission, is a speech disorder that can significantly compromise speech intelligibility. Cleft palate, previously repaired cleft palate and submucous cleft palate are associated with VPI. Less commonly, patients may acquire it after adenoidectomy with or without tonsillectomy or as a result of neuromuscular dysfunction. Comprehensive evaluation by a VPI team includes medical assessment focusing on airway obstructive symptoms, perceptual speech analysis, MRI and instrumental assessment. Options for intervention include speech therapy, intraoral prosthetic devices and surgery. Surgical methods can be categorized as palatal, palatopharyngeal or pharyngeal procedures. Each surgical approach has its strengths and limitations. Oro-maxillofacial surgeons are increasingly involved in the referral, evaluation, and treatment of velopharyngeal function. Therefore, understanding of physiology, anatomic structures, evaluation and treatment protocols in VPI is very important. This article presents protocol for evaluation of velopharyngeal function with a focus on indications for surgical interventions.

  • PDF

Recent advances in genetic studies of stuttering

  • Kang, Changsoo
    • Journal of Genetic Medicine
    • /
    • v.12 no.1
    • /
    • pp.19-24
    • /
    • 2015
  • Speech and language are uniquely human-specific traits, which contributed to humans becoming the predominant species on earth. Disruptions in the human speech and language function may result in diverse disorders. These include stuttering, aphasia, articulation disorder, spasmodic dysphonia, verbal dyspraxia, dyslexia and specific language impairment. Among these disorders, stuttering is the most common speech disorder characterized by disruptions in the normal flow of speech. Twin, adoption, and family studies have suggested that genetic factors are involved in susceptibility to stuttering. For several decades, multiple genetic studies including linkage analysis were performed to connect causative gene to stuttering, and several genetic studies have revealed the association of specific gene mutation with stuttering. One notable genetic discovery came from the genetic studies in the consanguineous Pakistani families. These studies suggested that mutations in the lysosomal enzyme-targeting pathway genes (GNPTAB, GNPTG and NAPGA) are associated with non-syndromic persistent stuttering. Although these studies have revealed some clues in understanding the genetic causes of stuttering, only a small fraction of patients are affected by these genes. In this study, we summarize recent advances and future challenges in an effort to understand genetic causes underlying stuttering.

Multi-Emotion Regression Model for Recognizing Inherent Emotions in Speech Data (음성 데이터의 내재된 감정인식을 위한 다중 감정 회귀 모델)

  • Moung Ho Yi;Myung Jin Lim;Ju Hyun Shin
    • Smart Media Journal
    • /
    • v.12 no.9
    • /
    • pp.81-88
    • /
    • 2023
  • Recently, communication through online is increasing due to the spread of non-face-to-face services due to COVID-19. In non-face-to-face situations, the other person's opinions and emotions are recognized through modalities such as text, speech, and images. Currently, research on multimodal emotion recognition that combines various modalities is actively underway. Among them, emotion recognition using speech data is attracting attention as a means of understanding emotions through sound and language information, but most of the time, emotions are recognized using a single speech feature value. However, because a variety of emotions exist in a complex manner in a conversation, a method for recognizing multiple emotions is needed. Therefore, in this paper, we propose a multi-emotion regression model that extracts feature vectors after preprocessing speech data to recognize complex, inherent emotions and takes into account the passage of time.

Phoneme distribution and phonological processes of orthographic and pronounced phrasal words in light of syllable structure in the Seoul Corpus (음절구조로 본 서울코퍼스의 글 어절과 말 어절의 음소분포와 음운변동)

  • Yang, Byunggon
    • Phonetics and Speech Sciences
    • /
    • v.8 no.3
    • /
    • pp.1-9
    • /
    • 2016
  • This paper investigated the phoneme distribution and phonological processes of orthographic and pronounced phrasal words in light of syllable structure in the Seoul Corpus in order to provide linguists and phoneticians with a clearer understanding of the Korean language system. To achieve the goal, the phrasal words were extracted from the transcribed label scripts of the Seoul Corpus using Praat. Following this, the onsets, peaks, codas and syllable types of the phrasal words were analyzed using an R script. Results revealed that k0 was most frequently used as an onset in both orthographic and pronounced phrasal words. Also, aa was the most favored vowel in the Korean syllable peak with fewer phonological processes in its pronounced form. The total proportion of all diphthongs according to the frequency of the peaks in the orthographic phrasal words was 8.8%, which was almost double those found in the pronounced phrasal words. For the codas, nn accounted for 34.4% of the total pronounced phrasal words and was the varied form. From syllable type classification of the Corpus, CV appeared to be the most frequent type followed by CVC, V, and VC from the orthographic forms. Overall, the onsets were more prevalent in the pronunciation more than the codas. From the results, this paper concluded that an analysis of phoneme distribution and phonological processes in light of syllable structure can contribute greatly to the understanding of the phonology of spoken Korean.