• Title/Summary/Keyword: Speech Perception

Search Result 397, Processing Time 0.033 seconds

The perception of clear and casual English speech under different speed conditions (다른 발화 속도의 또렷한 음성과 대화체로 발화한 영어문장 인지)

  • Yi, So Pae
    • Phonetics and Speech Sciences
    • /
    • v.10 no.2
    • /
    • pp.33-37
    • /
    • 2018
  • Korean students with much exposure to the relatively slow and clear speech used in most English classes in Korea can be expected to have difficulty understanding the casual style that is common in the everyday speech of English speakers. This research attempted to investigate an effective way to utilize casual speech in English education, by exploring the way different speech styles (clear vs. casual) affect Korean learners' comprehension of spoken English. Twenty Korean university students and two native speakers of English participated in a listening session. The English utterances were produced in different speech styles (clear slow, casual slow, clear fast, and casual fast). The Korean students were divided into two groups by English proficiency level. The results showed that the Korean students achieved 69.4% comprehension accuracy, while the native speakers of English demonstrated almost perfect results. The Korean students (especially the low-proficiency group) had more problems perceiving function words than they did perceiving content words. Responding to the different speech styles, the high-proficiency group had more difficulty listening to utterances with phonological variation than they did listening to utterances produced at a faster speed. The low-proficiency group, however, struggled with utterances produced at a faster speed more than they did with utterances with phonological variation. The pedagogical implications of the results are discussed in the concluding section.

Effects of attention on the perception of L2 phonetic contrast

  • Lee, Hyunjung
    • Phonetics and Speech Sciences
    • /
    • v.6 no.4
    • /
    • pp.47-52
    • /
    • 2014
  • This study investigated how the degree of attention modulates English learners' perception of Korean stop contrasts. The contributions of VOT and F0 in perceiving Korean stops were examined while availability of attentional resources was manipulated using a dual-task paradigm. Results demonstrated the attentional modulation in the use of VOT, but not in F0: under less attention, the contribution of VOT to the perception of aspirated stops decreased, whereas that of lenis stops increased, which suggests more native-like performance. This implies that the role of attention in perceiving non-native contrasts might differ depending on how equivalent the acoustic and perceptual cues are between L1 and target L2 contrasts.

Analysis of Lexical Effect on Spoken Word Recognition Test (한국어 단음절 낱말 인식에 미치는 어휘적 특성의 영향)

  • Yoon, Mi-Sun;Yi, Bong-Won
    • MALSORI
    • /
    • no.54
    • /
    • pp.15-26
    • /
    • 2005
  • The aim of this paper was to analyze the lexical effects on spoken word recognition of Korean monosyllabic word. The lexical factors chosen in this paper was frequency, density and lexical familiarity of words. Result of the analysis was as follows; frequency was the significant factor to predict spoken word recognition score of monosyllabic word. The other factors were not significant. This result suggest that word frequency should be considered in speech perception test.

  • PDF

Some effects of audio-visual speech in perceiving Korean

  • Kim, Jee-Sun;Davis, Chris
    • Annual Conference on Human and Language Technology
    • /
    • 1999.10e
    • /
    • pp.335-342
    • /
    • 1999
  • The experiments reported here investigated whether seeing a speaker's face (visible speech) affects the perception and memory of Korean speech sounds. In order to exclude the possibility of top-down, knowledge-based influences on perception and memory, the experiments tested people with no knowledge of Korean. The first experiment examined whether visible speech (Auditory and Visual - AV) assists English native speakers (with no knowledge of Korean) in the detection of a syllable within a Korean speech phrase. It was found that a syllable was more likely to be detected within a phrase when the participants could see the speaker's face. The second experiment investigated whether English native speakers' judgments about the duration of a Korean phrase would be affected by visible speech. It was found that in the AV condition participant's estimates of phrase duration were highly correlated with the actual durations whereas those in the AO condition were not. The results are discussed with respect to the benefits of communication with multimodal information and future applications.

  • PDF

A study on the perception of Korean phonation types by Aymara subjects (아이마라어 화자들의 한국어 발성유형 인지 연구)

  • Park, Hansang
    • Phonetics and Speech Sciences
    • /
    • v.8 no.4
    • /
    • pp.49-61
    • /
    • 2016
  • The present study investigates the perception of Korean phonation types by native speakers of Aymara. Perception tests were conducted on two sets of Korean speech materials to determine correspondence between Korean and Aymara 3-way contrasts and to find out which of the consonantal and vocalic part of the syllable is more influential in the perception of Korean phonation types. A set of manipulated stimuli, as well as a set of 12 spontaneous words, were prepared for the tests. The first syllable of the 12 Korean bisyllabic words of 3 series of phonation types(Lenis, Aspirated, and Fortis) in 4 places of articulation were split into consonantal and vocalic parts. And then the two parts were combined to form 9 tokens of CV sequences respectively for each place of articulation. Native speakers of Aymara were forced to match Korean stimuli with one of the 15 Aymara words which represent 3 series of consonant types(plain, aspirated, and ejective) in 5 places of articulation(bilabial, alveolar, palatal, velar, and uvular). Results showed that the consonantal part is more influential than the vocalic part to the Aymara subjects' perception of Korean phonation types when the consonantal part is Aspirated in its phonation type, but the vocalic part is more influential than the consonantal part when the consonantal part is Lenis or Fortis in its phonation type. Response analysis showed that Aymara subjects tend to match Korean stops to Aymara ones in such a way that Lenis corresponds to aspirated, Aspirated to aspirated, and Fortis to plain.

Acoustic correlates of prosodic prominence in conversational speech of American English, as perceived by ordinary listeners

  • Mo, Yoon-Sook
    • Phonetics and Speech Sciences
    • /
    • v.3 no.3
    • /
    • pp.19-26
    • /
    • 2011
  • Previous laboratory studies have shown that prosodic structures are encoded in the modulations of phonetic patterns of speech including suprasegmental as well as segmental features. Drawing on a prosodically annotated large-scale speech data from the Buckeye corpus of conversational speech of American English, the current study first evaluated the reliability of prosody annotation by a large number of ordinary listeners and later examined whether and how prosodic prominence influences the phonetic realization of multiple acoustic parameters in everyday conversational speech. The results showed that all the measures of acoustic parameters including pitch, loudness, duration, and spectral balance are increased when heard as prominent. These findings suggest that prosodic prominence enhances the phonetic characteristics of the acoustic parameters. The results also showed that the degree of phonetic enhancement vary depending on the types of the acoustic parameters. With respect to the formant structure, the findings from the present study more consistently support Sonority Expansion Hypothesis than Hyperarticulation Hypothesis, showing that the lexically stressed vowels are hyperarticulated only when hyperarticulation does not interfere with sonority expansion. Taken all into account, the present study showed that prosodic prominence modulates the phonetic realization of the acoustic parameters to the direction of the phonetic strengthening in everyday conversational speech and ordinary listeners are attentive to such phonetic variation associated with prosody in speech perception. However, the present study also showed that in everyday conversational speech there is no single dominant acoustic measure signaling prosodic prominence and listeners must attend to such small acoustic variation or integrate acoustic information from multiple acoustic parameters in prosody perception.

  • PDF

Nonlinear Interaction between Consonant and Vowel Features in Korean Syllable Perception (한국어 단음절에서 자음과 모음 자질의 비선형적 지각)

  • Bae, Moon-Jung
    • Phonetics and Speech Sciences
    • /
    • v.1 no.4
    • /
    • pp.29-38
    • /
    • 2009
  • This study investigated the interaction between consonants and vowels in Korean syllable perception using a speeded classification task (Garner, 1978). Experiment 1 examined whether listeners analytically perceive the component phonemes in CV monosyllables when classification is based on the component phonemes (a consonant or a vowel) and observed a significant redundancy gain and a Garner interference effect. These results imply that the perception of the component phonemes in a CV syllable is not linear. Experiment 2 examined the further relation between consonants and vowels at a subphonemic level comparing classification times based on glottal features (aspiration and lax), on place of articulation features (labial and coronal), and on vowel features (front and back). Across all feature classifications, there were significant but asymmetric interference effects. Glottal feature.based classification showed the least amount of interference effect, while vowel feature.based classification showed moderate interference, and place of articulation feature-based classification showed the most interference. These results show that glottal features are more independent to vowels, but place features are more dependent to vowels in syllable perception. To examine the three-way interaction among glottal, place of articulation, and vowel features, Experiment 3 featured a modified Garner task. The outcome of this experiment indicated that glottal consonant features are independent to both the place of articulation and vowel features, but the place of articulation features are dependent to glottal and vowel features. These results were interpreted to show that speech perception is not abstract and discrete, but nonlinear, and that the perception of features corresponds to the hierarchical organization of articulatory features which is suggested in nonlinear phonology (Clements, 1991; Browman and Goldstein, 1989).

  • PDF

Changes of Acoustic Reflex Thresholds and Speech-In-Noise Perception Using Personal Listening Device Under Subway Interior Noise (지하철 실내 소음 하에서 음향기기 사용에 따른 음향반사역치와 소음 속 어음인지 변화)

  • Han, Woojae;Chun, Hyungi;Ma, Sunmi
    • The Journal of the Acoustical Society of Korea
    • /
    • v.34 no.2
    • /
    • pp.138-145
    • /
    • 2015
  • Although it is well-known that environmental noise can lead to hearing loss in individuals, the true extent of subway noise effects in the general population remains poorly understood. The purpose of the present study is to see changes of acoustic reflex thresholds and speech perception scores when passengers listen to music presented from their personal listening device in the subway. Forty subjects with normal hearing participated being divided into two groups, experimental and control groups. As a baseline, all subjects were measured by acoustic reflex thresholds in five test frequencies and Korean speech perception in noise (KSPIN) test at 0 and -5 dB SNR. In the experiment, the control group read newspaper or magazine in the subway noise, whereas the experimental group listened to music presented from their smartphone under the subway noise through speakers at 73.45 dBA for 60 min. After completing the experiment, two groups also conducted both acoustic reflex thresholds and KSPIN tests in the same condition as the baseline. The results showed that there was a significant difference of correct percent in speech-in-noise test between experimental and control groups and of that between two signal-to-noise ratios, which means the double noise exposure of 60 min might cause some degradation of speech perception when noise increases compared to only subway noise condition that was not statistically significant difference. We concluded that a risk of some degraded speech perception ability would be expected when passengers have a habit of listening to music in the subway noisy situation for a long duration.

Speech Perception and Language Outcome in Congenitally Deaf Children Receiving Cochlear Implants in the First Year of Life (생후 12개월 이전에 인공와우이식을 받은 선천성 농 아동의 말지각과 언어능력)

  • Jeong, Sung Wook;Seo, Ji-Won;Boo, Sung-Hyun;Kim, Lee-Suk
    • Korean Journal of Otorhinolaryngology-Head and Neck Surgery
    • /
    • v.61 no.12
    • /
    • pp.644-649
    • /
    • 2018
  • Background and Objectives The objective of this study was to examine the effect of early cochlear implantation (CI) performed in infants less than 12 months of age. Subjects and Method Twenty-five children who received their first CI before 12 months of age were included in this study (infant group). The speech perception and language outcomes of these children were compared with those of 14 children who received their first CI between 13 and 24 months of age (older group). All children received sequential bilateral CI with the inter-stage interval of less than 2 years. Speech perception was measured using Categories of Auditory Performance, monosyllabic word test and sentence test, and language ability was measured using Sequenced Language Scale for Infants, Preschool Receptive-Expressive Language Scale, or Receptive & Expressive Vocavulary Test, depending on the age at the time of testing. Results There were no significant differences in speech perception abilities between the infant group and the older group. The mean expressive language score of infant group was higher than that of the older group, but the difference was not statistically significant. However, the receptive language score of infant group was significantly higher than that of the older group. Conclusion Children who received CI before 12 months of age achieved better receptive language ability than those who received it after 12 months of age. Thus CI should be performed as early as before 12 months of age to achieve better language ability.

A simulation study of speech perception enhancement for cochlear implant patients using companding in noisy environment (잡음 환경에서 압신을 이용한 인공 와우 환자의 언어 인지 향상 시뮬레이션 연구)

  • Lee Young-Woo;Ji Yoon-Sang;Lee Jong-Shil;Kim In-Young;Kim Sun-I.;Hong Sung-Hwa;Lee Sang-Min
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.5 s.311
    • /
    • pp.79-87
    • /
    • 2006
  • In this study, we evaluated the performance of a companding strategy as a preprocessing for speech enhancement and noise reduction. The proposed algorithm is based on two tone suppression that is human's hearing characteristics. This algorithm enhances spectral peak of speech signal and reduces background noise, however it has tradeoff characteristics between speech distortion and noise reduction due to limited channel number and nonlinear block. Therefore, we designed two different companding structures that have relative characteristics of noise reduction and speech distortion and found suitable companding structures by difference of individual speech perception ability in noise environment. Thus we proposed speech perception enhancement of cochlear implant user in noise environment with low SNR. The performance of the proposed algorithm was evaluated through 5 normal hearing listeners using noise band simulation. Improvement of speech perception was observed for all subjects and each subject preferred the different type of companding structure.