• Title/Summary/Keyword: Utterance

Search Result 382, Processing Time 0.026 seconds

Variance characteristics of speaking fundamental frequency and vocal intensity depending on utterance conditions (발화조건에 따른 기본주파수 및 음성강도 변동의 특징)

  • Lee, Moo-Kyung
    • Phonetics and Speech Sciences
    • /
    • v.4 no.1
    • /
    • pp.111-118
    • /
    • 2012
  • The purpose of this study was to characterize and determine variances of speaking fundamental frequency and vocal intensity depending on gender and three utterance conditions (spontaneous speech, reading, and counting). A total of 65 undergraduate students (32 male students, 33 female students) attending universities in Daegu, South Korea participated in this study. The subjects were all in their 20s. This study used KayPENTAX's Visi-Pitch IV (Model 3950) to measure the variances of speaking fundamental frequency (SFF0) and vocal intensity (VI). As a result, this study came to the following conclusions. First, it was found that both males and females showed no significant difference in SFF0 and vocal intensity among three utterance conditions. Second, this study sought to analyze differences in the variances of SFF0 between males and females. As a result, it was found that females showed significantly higher levels of four measured variances (SFF0 $SD^{**}$, SFF0 $range^{***}$, Min $SFF0^{***}$ and Max $SFF0^{***}$) than males on spontaneous speech. However, it was found that there was no significant difference between males and females in SFF0 range on reading or in SFF0 SD and SFF0 range on counting. It was found that there was no significant difference between males and females in the level of measured variances of vocal intensity depending on utterance conditions. Finally, this study made a comparison and analysis on differences in the variances of SFF0 and vocal intensity among utterance conditions. As a result, it was found that all the measured variances of SFF0 in males were most significantly reduced depending upon spontaneous speech which was followed by reading and counting respectively (SFF0 SD: p<.001, SFF0 range: p<.05, Max SFF0: p<.05). Females however, show no significant difference in the measured variances of SFF0 depending upon three utterance conditions. It was also found that the measured variances of vocal intensity in females were most significantly reduced depending on spontaneous speech that was followed by reading and counting (VI SD: p<.001, VI range: p<.001, Min VI: p<.01 Max VI: p<.05), while males showed no significant difference in the measured variances of vocal intensity depending on three utterance conditions. In sum, these findings suggest that variances of SFF0 in males are affected by three utterance conditions, while variances of vocal intensity in females are affected by three utterance conditions.

A Method for Measuring Inter-Utterance Similarity Considering Various Linguistic Features (다양한 언어적 자질을 고려한 발화간 유사도 측정 방법)

  • Lee, Yeon-Su;Shin, Joong-Hwi;Hong, Gum-Won;Song, Young-In;Lee, Do-Gil;Rim, Hae-Chang
    • The Journal of the Acoustical Society of Korea
    • /
    • v.28 no.1
    • /
    • pp.61-69
    • /
    • 2009
  • This paper presents an improved method measuring inter-utterance similarity in an example-based dialogue system, which searches the most similar utterance in a dialogue database to generate a response to a given user utterance. Unlike general inter-sentence similarity measures, the inter-utterance similarity measure for example-based dialogue system should consider not only word distribution but also various linguistic features, such as affirmation/negation, tense, modality, sentence type, which affects the natural conversation. However, previous approaches do not sufficiently reflect these features. This paper proposes a new utterance similarity measure by analyzing and reflecting various linguistic features to improve performance in accuracy. Also, by considering substitutability of the features, the proposed method can utilize limited number of examples. Experimental results show that the proposed method achieves 10%p improvement in accuracy compared to the previous method.

A realization of pauses in utterance across speech style, gender, and generation (과제, 성별, 세대에 따른 휴지의 실현 양상 연구)

  • Yoo, Doyoung;Shin, Jiyoung
    • Phonetics and Speech Sciences
    • /
    • v.11 no.2
    • /
    • pp.33-44
    • /
    • 2019
  • This paper dealt with how realization of pauses in utterance is affected by speech style, gender, and generation. For this purpose, we analyzed the frequency and duration of pauses. Pauses were categorized into four types: pause with breath, pause with no breath, utterance medial pause, and utterance final pause. Forty-eight subjects living in Seoul were chosen from the Korean Standard Speech Database. All subjects engaged in reading and spontaneous speech, through which we could also compare the realization between the two speech styles. The results showed that utterance final pauses had longer durations than utterance medial pauses. It means that utterance final pause has a function that signals the end of an utterance to the audience. For difference between tasks, spontaneous speech had longer and more frequent pauses because of cognitive reasons. With regard to gender variables, women produced shorter and less frequent pauses. For male speakers, the duration of pauses with breath was significantly longer. Finally, for generation variable, older speakers produced more frequent pauses. In addition, the results showed several interaction effects. Male speakers produced longer pauses, but this gender effect was more prominent at the utterance final position.

Utterance Verification Using Anti-models Based on Neighborhood Information (이웃 정보에 기초한 반모델을 이용한 발화 검증)

  • Yun, Young-Sun
    • MALSORI
    • /
    • no.67
    • /
    • pp.79-102
    • /
    • 2008
  • In this paper, we investigate the relation between Bayes factor and likelihood ratio test (LRT) approaches and apply the neighborhood information of Bayes factor to building an alternate hypothesis model of the LRT system. To consider the neighborhood approaches, we contemplate a distance measure between models and algorithms to be applied. We also evaluate several methods to improve performance of utterance verification using neighborhood information. Among these methods, the system which adopts anti-models built by collecting mixtures of neighborhood models obtains maximum error rate reduction of 17% compared to the baseline, linear and weighted combination of neighborhood models.

  • PDF

Weighted Finite State Transducer-Based Endpoint Detection Using Probabilistic Decision Logic

  • Chung, Hoon;Lee, Sung Joo;Lee, Yun Keun
    • ETRI Journal
    • /
    • v.36 no.5
    • /
    • pp.714-720
    • /
    • 2014
  • In this paper, we propose the use of data-driven probabilistic utterance-level decision logic to improve Weighted Finite State Transducer (WFST)-based endpoint detection. In general, endpoint detection is dealt with using two cascaded decision processes. The first process is frame-level speech/non-speech classification based on statistical hypothesis testing, and the second process is a heuristic-knowledge-based utterance-level speech boundary decision. To handle these two processes within a unified framework, we propose a WFST-based approach. However, a WFST-based approach has the same limitations as conventional approaches in that the utterance-level decision is based on heuristic knowledge and the decision parameters are tuned sequentially. Therefore, to obtain decision knowledge from a speech corpus and optimize the parameters at the same time, we propose the use of data-driven probabilistic utterance-level decision logic. The proposed method reduces the average detection failure rate by about 14% for various noisy-speech corpora collected for an endpoint detection evaluation.

Modality Classification for an Example-Based Dialogue System (예제 기반 대화 시스템을 위한 양태 분류)

  • Kim, Min-Jeong;Hong, Gum-Won;Song, Young-In;Lee, Yeon-Soo;Lee, Do-Gil;Rim, Hae-Chang
    • MALSORI
    • /
    • v.68
    • /
    • pp.75-93
    • /
    • 2008
  • An example-based dialogue system tries to utilize many pairs which are stored in a dialogue database. The most important part of the example-based dialogue system is to find the most similar utterance to user's input utterance. Modality, which is characterized as conveying the speaker's involvement in the propositional content of a given utterance, is one of the core sentence features. For example, the sentence "I want to go to school." has a modality of hope. In this paper, we have proposed a modality classification system which can predict sentence modality in order to improve the performance of example-based dialogue systems. We also define a modality tag set for a dialogue system, and validate this tag set using a rule-based modality classification system. Experimental results show that our modality tag set and modality classification system improve the performance of an example-based dialogue system.

  • PDF

Utterance Verification and Substitution Error Correction In Korean Connected Digit Recognition (한국어 연결숫자 인식에서의 발화 검증과 대체오류 수정)

  • Jung Du Kyung;Song Hwa Jeon;Jung Ho-Young;Kim Hyung Soon
    • MALSORI
    • /
    • no.45
    • /
    • pp.79-91
    • /
    • 2003
  • Utterance verification aims at rejecting both out-of-vocabulary (OOV) utterances and low-confidence-scored in-vocabulary (IV) utterances. For utterance verification on Korean connected digit recognition task, we investigate several methods to construct filler and anti-digit models. In particular, we propose a substitution error correction method based on 2-best decoding results. In this method, when 1st candidate is rejected, 2nd candidate is selected if it is accepted by a specific hypothesis test, instead of simply rejecting the 1st one. Experimental results show that the proposed method outperforms the conventional log likelihood ratio (LLR) test method.

  • PDF

A Study on the Speech Intelligibility of Voice Disordered Patients according to the Severity and Utterance Level (음성장애의 중증도와 발화 수준에 따른 말 명료도의 변화 연구)

  • Pyo, Hwa-Young
    • Speech Sciences
    • /
    • v.15 no.2
    • /
    • pp.101-110
    • /
    • 2008
  • The purpose of this study was to investigate the speech intelligibility of voice disordered patients when we consider the severity and utterance level as variables. Based on the severity level, 12 patients were divided into three groups, G1, G2, and G3 group, respectively. Words, phrases and sentences produced by the speakers were judged by four listeners with normal hearing, and we compared the intelligibility scores of the three groups. As a result, the speech intelligibility was decreased as the severity level was increased, and the difference was statistically significant. However, the mean difference among words, phrases and sentences was not significant, and the variation of intelligibility according to the utterance level was not under the regular rules.

  • PDF

Focal Parts of Utterance in Busan Korean

  • Cho, Yong-Hyung
    • Speech Sciences
    • /
    • v.9 no.4
    • /
    • pp.149-163
    • /
    • 2002
  • Focal parts of utterance can be determined by new/contrastive information, a focus particle, a contrastive topic marker, or a nominative case marker in Busan Korean. Among these factors, new or contrastive information is the most important element in determining the intonational nucleus of an utterance. However, unlike Seoul Korean, when a focus particle, a topic marker, or a case marker contributes to the placement of the most prominent peak of an utterance, the peak is on the noun to which they are attached. Moreover, the case marker-ga shows more prominent pitch on the preceding noun than the noun followed by the topic marker-nun when-ga is used as emphatic or contrastive. This is one of the major problems for Busan Korean users in commanding natural and fluent Seoul Korean intonation even if they use standard written form of Seoul Korean in their speech.

  • PDF

Confidence Measure for Utterance Verification in Noisy Environments (잡음 환경에서의 인식 거부 성능 향상을 위한 신뢰 척도)

  • Park, Jeong-Sik;Oh, Yung-Hwan
    • Proceedings of the KSPS conference
    • /
    • 2006.11a
    • /
    • pp.3-6
    • /
    • 2006
  • This paper proposes a confidence measure employed for utterance verification in noisy environments. Most of conventional approaches estimate the proper threshold of confidence measure and apply the value to utterance rejection in recognition process. As such, their performance may degrade for noisy speech since the threshold can be changed in noisy environments. This paper presents further robust confidence measure based on the multi-pass confidence measure. The isolated word recognition based experimental results demonstrate that the proposed method outperforms conventional approaches as utterance verifier.

  • PDF