• 제목/요약/키워드: Phonetic Approach

검색결과 78건 처리시간 0.023초

Phonetic Approach or Phonological Approach: Syntax-prosody Interface in Seoul Korean

  • Utsugi, Akira
    • 음성과학
    • /
    • 제11권4호
    • /
    • pp.207-221
    • /
    • 2004
  • There are two different approaches in studying mapping between syntactic structure and prosody, the 'phonetic approach' and the 'phonological approach'. An experiment to examine which approach is more valid was conducted. In the experiment, syntactically ambiguous Seoul Korean sentences in each of which a noun immediately after an adjective starts with either an H-segment (a segment which triggers the AP-initial H tone) or an L-segment (a segment other than H-segments) were recorded. by 3 Seoul Korean speakers. The F0 values in the syllables containing the consonants in question were then measured. The results show that interaction between the segment type and the branching structure is statistically significant. which suggests that it is difficult to use the phonetic .approach to generalize the relationship between syntax and prosody. Thus, it is concluded that the phonological approach is more valid.

  • PDF

Acoustic-Phonetic Phenotypes in Pediatric Speech Disorders;An Interdisciplinary Approach

  • Bunnell, H. Timothy
    • 대한음성학회:학술대회논문집
    • /
    • 대한음성학회 2006년도 추계학술대회 발표논문집
    • /
    • pp.31-36
    • /
    • 2006
  • Research in the Center for Pediatric Auditory and Speech Sciences (CPASS) is attempting to characterize or phenotype children with speech delays based on acoustic-phonetic evidence and relate those phenotypes to chromosome loci believed to be related to language and speech. To achieve this goal we have adopted a highly interdisciplinary approach that merges fields as diverse as automatic speech recognition, human genetics, neuroscience, epidemiology, and speech-language pathology. In this presentation I will trace the background of this project, and the rationale for our approach. Analyses based on a large amount of speech recorded from 18 children with speech delays will be presented to illustrate the approach we will be taking to characterize the acoustic phonetic properties of disordered speech in young children. The ultimate goal of our work is to develop non-invasive and objective measures of speech development that can be used to better identify which children with apparent speech delays are most in need of, or would receive the most benefit from the delivery of therapeutic services.

  • PDF

A Feature-based Approach to English Phonetic Mastery --Cognitive and/or Physical--

  • Takashi Shimaoka
    • 대한음성학회:학술대회논문집
    • /
    • 대한음성학회 1996년도 10월 학술대회지
    • /
    • pp.349-354
    • /
    • 1996
  • The phonetic mastery of English has been considered next to impossible to many non-native speakers of English, including even some teachers of English. This paper takes issue with this phonetic problem of second language acquisition and proposes that combination of cognitive and physical approaches can help master English faster and more easily.

  • PDF

음소변동규칙의 적합도 조정을 통한 연속음성인식 성능향상 (Improving the Performance of the Continuous Speech Recognition by Estimating Likelihoods of the Phonetic Rules)

  • 나민수;정민화
    • 대한음성학회:학술대회논문집
    • /
    • 대한음성학회 2006년도 추계학술대회 발표논문집
    • /
    • pp.80-83
    • /
    • 2006
  • The purpose of this paper is to build a pronunciation lexicon with estimated likelihoods of the phonetic rules based on the phonetic realizations and therefore to improve the performance of CSR using the dictionary. In the baseline system, the phonetic rules and their application probabilities are defined with the knowledge of Korean phonology and experimental tuning. The advantage of this approach is to implement the phonetic rules easily and to get stable results on general domains. However, a possible drawback of this method is that it is hard to reflect characteristics of the phonetic realizations on a specific domain. In order to make the system reflect phonetic realizations, the likelihood of phonetic rules is reestimated based on the statistics of the realized phonemes using a forced-alignment method. In our experiment, we generates new lexica which include pronunciation variants created by reestimated phonetic rules and its performance is tested with 12 Gaussian mixture HMMs and back-off bigrams. The proposed method reduced the WER by 0.42%.

  • PDF

Possibililty of the Rough Set Approach in Phonetic Distinctions

  • Lee, Jae-Ik;Kim, Jeong-Kuk;Jo, Heung-Kuk;Suh, Kap-Sun
    • 한국음향학회:학술대회논문집
    • /
    • 한국음향학회 1996년도 영남지부 학술발표회 논문집 Acoustic Society of Korean Youngnam Chapter Symposium Proceedings
    • /
    • pp.66-69
    • /
    • 1996
  • The goal of automatic speech recognition is to study techniques and systems that enable agents such that computers and robots to accept speech input. However, this paper does not provide a concrete technology in speech recognition but propose a possible mathematical tools to be employed in that area. We introduce rough set theory and suggest the possibility of the rough set approach in phonetic distinctions.

  • PDF

Korean Fortis Consonants and Post Obstruent Tensifcation: A Cognitive Approach

  • Ko, Eon-Suk
    • 대한음성학회:학술대회논문집
    • /
    • 대한음성학회 1996년도 10월 학술대회지
    • /
    • pp.482-487
    • /
    • 1996
  • Korean fortis consonant is not included in the consonantal inventory, but a result of phonetic implementation at the phonetic level, P. With the framework of Cognitive Phonology, a construction of Post Obstruent Tensification is proposed in such a way that rule-ordering is eliminated. This enables us to overcome methodological problems raised in former analyses of fortis under geminate hypothesis, and give a uniform account for three categories of fortis consonants. By assuming extrasyllabicity of verb-stem-final, consonant neutralization of fortis in the coda position is explained by the invisibility at the P-level. and, therefore. modification of Coda Neutralization rule is called for.

  • PDF

일본어 합성기에서 악센트 정보가 결합된 발음기호를 이용한 Break 예측 방법 (Break Predicting Methods Using Phonetic Symbols Combined with Accents Information in a Japanese Speech Synthesizer)

  • 나덕수;이종석;김종국;배명진
    • 대한음성학회지:말소리
    • /
    • 제62호
    • /
    • pp.69-84
    • /
    • 2007
  • Japanese is a language having intonations, which are indicated by the relative differences in pitch heights and the accentual phrases (APs) are placed according to the changes of the accents while a break occurs on a boundary of the APs. Although a break can be predicted by using J-ToBI, which is a rule-based or statistical approach, it is very difficult to predict a break exactly due to the flexibility. Therefore, in this paper, a method which can enhance the quality of synthesized speech by reducing the errors in predicting break indices (BI), are proposed. The method is to use a new definition for the phonetic symbols, which combine the phonetic values of Japanese words with the accents information. Since a stream of defined phonetic symbols includes the information on the changes in intonations, the BI can be easily predicted by dividing the intonation phrase (IP) into several APs. As a result of an experiment, the accuracy of break generations was 98 % and the proposed method contributed itself to enhance the naturalness of synthesized speeches.

  • PDF

Text-driven Speech Animation with Emotion Control

  • Chae, Wonseok;Kim, Yejin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • 제14권8호
    • /
    • pp.3473-3487
    • /
    • 2020
  • In this paper, we present a new approach to creating speech animation with emotional expressions using a small set of example models. To generate realistic facial animation, two example models called key visemes and expressions are used for lip-synchronization and facial expressions, respectively. The key visemes represent lip shapes of phonemes such as vowels and consonants while the key expressions represent basic emotions of a face. Our approach utilizes a text-to-speech (TTS) system to create a phonetic transcript for the speech animation. Based on a phonetic transcript, a sequence of speech animation is synthesized by interpolating the corresponding sequence of key visemes. Using an input parameter vector, the key expressions are blended by a method of scattered data interpolation. During the synthesizing process, an importance-based scheme is introduced to combine both lip-synchronization and facial expressions into one animation sequence in real time (over 120Hz). The proposed approach can be applied to diverse types of digital content and applications that use facial animation with high accuracy (over 90%) in speech recognition.