• Title/Summary/Keyword: Phonetic Approach

Search Result 78, Processing Time 0.021 seconds

Phonetic Approach or Phonological Approach: Syntax-prosody Interface in Seoul Korean

  • Utsugi, Akira
    • Speech Sciences
    • /
    • v.11 no.4
    • /
    • pp.207-221
    • /
    • 2004
  • There are two different approaches in studying mapping between syntactic structure and prosody, the 'phonetic approach' and the 'phonological approach'. An experiment to examine which approach is more valid was conducted. In the experiment, syntactically ambiguous Seoul Korean sentences in each of which a noun immediately after an adjective starts with either an H-segment (a segment which triggers the AP-initial H tone) or an L-segment (a segment other than H-segments) were recorded. by 3 Seoul Korean speakers. The F0 values in the syllables containing the consonants in question were then measured. The results show that interaction between the segment type and the branching structure is statistically significant. which suggests that it is difficult to use the phonetic .approach to generalize the relationship between syntax and prosody. Thus, it is concluded that the phonological approach is more valid.

  • PDF

Acoustic-Phonetic Phenotypes in Pediatric Speech Disorders;An Interdisciplinary Approach

  • Bunnell, H. Timothy
    • Proceedings of the KSPS conference
    • /
    • 2006.11a
    • /
    • pp.31-36
    • /
    • 2006
  • Research in the Center for Pediatric Auditory and Speech Sciences (CPASS) is attempting to characterize or phenotype children with speech delays based on acoustic-phonetic evidence and relate those phenotypes to chromosome loci believed to be related to language and speech. To achieve this goal we have adopted a highly interdisciplinary approach that merges fields as diverse as automatic speech recognition, human genetics, neuroscience, epidemiology, and speech-language pathology. In this presentation I will trace the background of this project, and the rationale for our approach. Analyses based on a large amount of speech recorded from 18 children with speech delays will be presented to illustrate the approach we will be taking to characterize the acoustic phonetic properties of disordered speech in young children. The ultimate goal of our work is to develop non-invasive and objective measures of speech development that can be used to better identify which children with apparent speech delays are most in need of, or would receive the most benefit from the delivery of therapeutic services.

  • PDF

A Feature-based Approach to English Phonetic Mastery --Cognitive and/or Physical--

  • Takashi Shimaoka
    • Proceedings of the KSPS conference
    • /
    • 1996.10a
    • /
    • pp.349-354
    • /
    • 1996
  • The phonetic mastery of English has been considered next to impossible to many non-native speakers of English, including even some teachers of English. This paper takes issue with this phonetic problem of second language acquisition and proposes that combination of cognitive and physical approaches can help master English faster and more easily.

  • PDF

Improving the Performance of the Continuous Speech Recognition by Estimating Likelihoods of the Phonetic Rules (음소변동규칙의 적합도 조정을 통한 연속음성인식 성능향상)

  • Na, Min-Soo;Chung, Min-Hwa
    • Proceedings of the KSPS conference
    • /
    • 2006.11a
    • /
    • pp.80-83
    • /
    • 2006
  • The purpose of this paper is to build a pronunciation lexicon with estimated likelihoods of the phonetic rules based on the phonetic realizations and therefore to improve the performance of CSR using the dictionary. In the baseline system, the phonetic rules and their application probabilities are defined with the knowledge of Korean phonology and experimental tuning. The advantage of this approach is to implement the phonetic rules easily and to get stable results on general domains. However, a possible drawback of this method is that it is hard to reflect characteristics of the phonetic realizations on a specific domain. In order to make the system reflect phonetic realizations, the likelihood of phonetic rules is reestimated based on the statistics of the realized phonemes using a forced-alignment method. In our experiment, we generates new lexica which include pronunciation variants created by reestimated phonetic rules and its performance is tested with 12 Gaussian mixture HMMs and back-off bigrams. The proposed method reduced the WER by 0.42%.

  • PDF

Possibililty of the Rough Set Approach in Phonetic Distinctions

  • Lee, Jae-Ik;Kim, Jeong-Kuk;Jo, Heung-Kuk;Suh, Kap-Sun
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • 1996.10a
    • /
    • pp.66-69
    • /
    • 1996
  • The goal of automatic speech recognition is to study techniques and systems that enable agents such that computers and robots to accept speech input. However, this paper does not provide a concrete technology in speech recognition but propose a possible mathematical tools to be employed in that area. We introduce rough set theory and suggest the possibility of the rough set approach in phonetic distinctions.

  • PDF

Korean Fortis Consonants and Post Obstruent Tensifcation: A Cognitive Approach

  • Ko, Eon-Suk
    • Proceedings of the KSPS conference
    • /
    • 1996.10a
    • /
    • pp.482-487
    • /
    • 1996
  • Korean fortis consonant is not included in the consonantal inventory, but a result of phonetic implementation at the phonetic level, P. With the framework of Cognitive Phonology, a construction of Post Obstruent Tensification is proposed in such a way that rule-ordering is eliminated. This enables us to overcome methodological problems raised in former analyses of fortis under geminate hypothesis, and give a uniform account for three categories of fortis consonants. By assuming extrasyllabicity of verb-stem-final, consonant neutralization of fortis in the coda position is explained by the invisibility at the P-level. and, therefore. modification of Coda Neutralization rule is called for.

  • PDF

Break Predicting Methods Using Phonetic Symbols Combined with Accents Information in a Japanese Speech Synthesizer (일본어 합성기에서 악센트 정보가 결합된 발음기호를 이용한 Break 예측 방법)

  • Na, Deok-Su;Lee, Jong-Seok;Kim, Jong-Kuk;Bae, Myung-Jin
    • MALSORI
    • /
    • no.62
    • /
    • pp.69-84
    • /
    • 2007
  • Japanese is a language having intonations, which are indicated by the relative differences in pitch heights and the accentual phrases (APs) are placed according to the changes of the accents while a break occurs on a boundary of the APs. Although a break can be predicted by using J-ToBI, which is a rule-based or statistical approach, it is very difficult to predict a break exactly due to the flexibility. Therefore, in this paper, a method which can enhance the quality of synthesized speech by reducing the errors in predicting break indices (BI), are proposed. The method is to use a new definition for the phonetic symbols, which combine the phonetic values of Japanese words with the accents information. Since a stream of defined phonetic symbols includes the information on the changes in intonations, the BI can be easily predicted by dividing the intonation phrase (IP) into several APs. As a result of an experiment, the accuracy of break generations was 98 % and the proposed method contributed itself to enhance the naturalness of synthesized speeches.

  • PDF

Text-driven Speech Animation with Emotion Control

  • Chae, Wonseok;Kim, Yejin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.8
    • /
    • pp.3473-3487
    • /
    • 2020
  • In this paper, we present a new approach to creating speech animation with emotional expressions using a small set of example models. To generate realistic facial animation, two example models called key visemes and expressions are used for lip-synchronization and facial expressions, respectively. The key visemes represent lip shapes of phonemes such as vowels and consonants while the key expressions represent basic emotions of a face. Our approach utilizes a text-to-speech (TTS) system to create a phonetic transcript for the speech animation. Based on a phonetic transcript, a sequence of speech animation is synthesized by interpolating the corresponding sequence of key visemes. Using an input parameter vector, the key expressions are blended by a method of scattered data interpolation. During the synthesizing process, an importance-based scheme is introduced to combine both lip-synchronization and facial expressions into one animation sequence in real time (over 120Hz). The proposed approach can be applied to diverse types of digital content and applications that use facial animation with high accuracy (over 90%) in speech recognition.