• Title/Summary/Keyword: English Vowel

Search Result 246, Processing Time 0.02 seconds

An Acoustical Study on the Syllable Structures of Korean Numeric Sounds (국어 숫자음의 음절구조에 대한 음향적 분석)

  • Yang, Byung-Gon
    • Proceedings of the KSPS conference
    • /
    • 2007.05a
    • /
    • pp.170-172
    • /
    • 2007
  • The purpose of this study was to examine the syllable structures of ten Korean numeric sounds produced by ten subjects of the same age. Each sound was normalized and divided into onset, vowel, and coda sections. Then, acoustical measurements of each syllable were done to compare the ten sounds. Results showed that there was not much deviation from the grand average duration and intensity for the majority of the sounds except the two diphthongal sounds on which their boundary points varied among the speakers. Some syllable boundaries were quite obvious while others were ambiguous. There seemed some tradeoff among the syllable components depending on their acoustic features.

  • PDF

Error Correction and Praat Script Tools for the Buckeye Corpus of Conversational Speech (벅아이 코퍼스 오류 수정과 코퍼스 활용을 위한 프랏 스크립트 툴)

  • Yoon, Kyu-Chul
    • Phonetics and Speech Sciences
    • /
    • v.4 no.1
    • /
    • pp.29-47
    • /
    • 2012
  • The purpose of this paper is to show how to convert the label files of the Buckeye Corpus of Spontaneous Speech [1] into Praat format and to introduce some of the Praat scripts that will enable linguists to study various aspects of spoken American English present in the corpus. During the conversion process, several types of errors were identified and corrected either manually or automatically by the use of scripts. The Praat script tools that have been developed can help extract from the corpus massive amounts of phonetic measures such as the VOT of plosives, the formants of vowels, word frequency information and speech rates that span several consecutive words. The script tools can extract additional information concerning the phonetic environment of the target words or allophones.

A Review of Timing Factors in Speech

  • Yun, Il-Sung
    • Speech Sciences
    • /
    • v.7 no.3
    • /
    • pp.87-98
    • /
    • 2000
  • Timing in speech is determined by many factors. In this paper, we introduce and discuss some factors that have generally been regarded as important in speech timing. They include stress, syllable structure, consonant insertion or deletion, tempo, lengthening at clause, phrase and word boundaries, preconsonantal vowel shortening, and compensation between segments or within phonological units (e.g., word, foot), compression due to the increase of syllables in word or foot level, etc. and each of them may playa crucial role in the structuring of speech timing in a language. But some of these timing factors must interact with each other rather than be independent and the effects of each factor on speech timing will vary from language to language. On the other hand, there could well be many other factors unknown so far. Finding out and investigating new timing factors and reinterpreting the already-known timing factors should enhance our understanding of timing structures in a given language or languages.

  • PDF

An Acoustic Study of the Stress and Intonational System in Lakhota: A Preliminary Report

  • Cho, Tae-Hong
    • Speech Sciences
    • /
    • v.13 no.4
    • /
    • pp.23-42
    • /
    • 2006
  • This paper reports a preliminary result of an acoustic study on the stress and intonational system in Lakhota, a native American language. It investigates how the stress and intonation in Lakhota are phonetically manifested; and how the stress interacts with other prosodic factors. The results preliminarily obtained from one native Lakhota speaker suggest that the primary cue of the stress is relatively high F0 which is often accompanied by higher intensity (for the vowel) and longer VOT (for aspirated stops). The results also indicate that stress is not reliably marked by duration. The stress system, however, interacts with the intonational pattern, such that, for example, intonational peak falls on the stressed syllable with a general pattern of L+H* and that it interacts with the boundary tone L%, resulting in mid tone utterance-finally. This paper can be viewed largely as a qualitative study on an understudied native American language, Lakhota and as forming a basis for further development of its stress and intonation system whose acoustic properties of its prosodic system have not been investigated before.

  • PDF

Korean Agraphia Subsequent to Right Hemispheric Lesion (우반구 손상 환자의 한글 실서증 특징)

  • Yoon, Ji-Hye;Shin, Ji-Cheol;Kim, Deog-Young;Suh, Mee-Kyung;Kim, Hyang-Hee
    • Speech Sciences
    • /
    • v.13 no.3
    • /
    • pp.121-132
    • /
    • 2006
  • In Hangeul, the graphemes of syllables are organized in horizontal, vertical and mixed (both horizontal & vertical) orientations, and the graphemic position of consonant(s) and vowel(s) within a each syllable needs to be maintained within a square pattern. We investigated the characteristics of writing errors of 9 stroke patients with right hemisphere (RH) lesions and compared it to the performances of 15 normal subjects. The subjects were asked to write to dictation of 90 Korean syllables. One of the interesting findings was that our patients manifested visuospatial errors which are not commonly observed in other language-speaking (e.g., English) patients due to the unique syllabic organizations of Korean writing system. The prominent errors in the RH group could be explained by the impaired RH which normally controls the visuospatial functions.

  • PDF

Lexical Status and the Degree of /l/-darkening

  • Ahn, Miyeon
    • Phonetics and Speech Sciences
    • /
    • v.7 no.3
    • /
    • pp.73-78
    • /
    • 2015
  • This study explores the degree of velarization of English word-final /l/ (i.e., /l/-darkness) according to the lexical status. Lexical status is defined as whether a speech stimulus is considered as a word or a non-word. We examined the temporal and spectral properties of word-final /l/ in terms of the duration and the frequency difference of F2-F1 values by varying the immediate pre-liquid vowels. The result showed that both temporal and spectral properties were contrastive across all vowel contexts in the way of real words having shorter [l] duration and low F2-F1 values, compared to non-words. That is, /l/ is more heavily velarized in words than in non-words, which suggests that lexical status whether language users encode the speech signal as a word or not is deeply involved in their speech production.

The Basic Study on making mono-phone for Korean Speech Recognition (한국어 음성 인식을 위한 mono-phone 구성의 기초 연구)

  • Hwang YoungSoo;Song Minsuck
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • autumn
    • /
    • pp.45-48
    • /
    • 2000
  • In the case of making large vocabulary speech recognition system, it is better to use the segment than the syllable or the word as the recognition unit. In this paper, we study on the basis of making mono-phone for Korean speech recognition. For experiments, we use the speech toolkit of OGI in U.S.A. The result shows that the recognition rate of :he case in which the diphthong is established as a single unit is superior to that of the case in which the diphthong is established as two units, i.e. a glide plus a vowel. And also, the recognition rate by the number of consonants is a little different.

  • PDF

Korean Phoneme Recognition Using Neural Networks (신경회로망 이용한 한국어 음소 인식)

  • 김동국;정차균;정홍
    • The Transactions of the Korean Institute of Electrical Engineers
    • /
    • v.40 no.4
    • /
    • pp.360-373
    • /
    • 1991
  • Since 70's, efficient speech recognition methods such as HMM or DTW have been introduced primarily for speaker dependent isolated words. These methods however have confronted with difficulties in recognizing continuous speech. Since early 80's, there has been a growing awareness that neural networks might be more appropriate for English and Japanese phoneme recognition using neural networks. Dealing with only a part of vowel or consonant set, Korean phoneme recognition still remains on the elementary level. In this light, we develop a system based on neural networks which can recognize major Korean phonemes. Through experiments using two neural networks, SOFM and TDNN, we obtained remarkable results. Especially in the case of using TDNN, the recognition rate was estimated about 93.78% for training data and 89.83% for test data.

Fundamental Acoustic Investigation of Korean Male 5 Monophthongs (한국 남성의 단모음 [아, 에, 이, 오, 우]에 대한 음향음성학적 기반연구)

  • Choi, Yae-Lin
    • The Journal of the Korea Contents Association
    • /
    • v.10 no.6
    • /
    • pp.373-377
    • /
    • 2010
  • Numerous quantitative and qualitative studies have already been published related to English vowels. However, only minimal amounts of studies based on the acoustic analysis of Korean vowels have been accomplished. The purpose of this study is to obtain sufficient quantitative data based on the acoustic aspects of Korean vowels produced by males between the ages of 20s and 30s. A total of 31 males in their 20s and 30s produced the five fundamental vowels /a, e, i, o, u/ by repeating each of them three times in the standard Korean dialect. Such speech productions were recorded with 'Cool edit' and F1, F2, F3, F4 were extracted through the MATLAB acoustic analysis program. Results indicated that the overall patterns of formants were similar to previous studies, except that the formant levels of F1 and F2 of the vowels produced in this study were generally lower than that in previous studies. Future studies need to focus on obtaining vowel data by considering other factors such as age and other speech materials.

Implementation of TTS Engine for Natural Voice (자연음 TTS(Text-To-Speech) 엔진 구현)

  • Cho Jung-Ho;Kim Tae-Eun;Lim Jae-Hwan
    • Journal of Digital Contents Society
    • /
    • v.4 no.2
    • /
    • pp.233-242
    • /
    • 2003
  • A TTS(Text-To-Speech) System is a computer-based system that should be able to read any text aloud. To output a natural voice, we need a general knowledge of language, a lot of time, and effort. Furthermore, the sound pattern of english has a variable pattern, which consists of phonemic and morphological analysis. It is very difficult to maintain consistency of pattern. To handle these problems, we present a system based on phonemic analysis for vowel and consonant. By analyzing phonological variations frequently found in spoken english, we have derived about phonemic contexts that would trigger the multilevel application of the corresponding phonological process, which consists of phonemic and allophonic rules. In conclusion, we have a rule data which consists of phoneme, and a engine which economize in system. The proposed system can use not only communication system, but also utilize office automation and so on.

  • PDF