• Title/Summary/Keyword: Dynamic speaker

Search Result 87, Processing Time 0.023 seconds

A Study on Will as Modal or Non-modal

  • Lee, Young Mi;Kang, Mun Koo
    • English Language & Literature Teaching
    • /
    • v.18 no.3
    • /
    • pp.175-190
    • /
    • 2012
  • The purpose of this article is to explain the meanings and uses of the English auxiliaries will morpho-syntactically, and answer the question of whether will is a tense auxiliary or a modal one. Some writers even exclude will completely from the semantics of the modal auxiliaries. They argue that the semantics of will is fundamentally non-modal and has only a few modal-like uses. There are some people who treat will to be semantically separate from the other modal auxiliaries. In the light of modal will, the semantics of will basically remains anchored in volition because the lack of required speaker subjectivity, but has undergone so much semantic bleaching that it may also express future time without volition. On the other hand, the semantics of will in the exclusionist view is erroneous and that its semantics is in fact closely related to the semantics of the other modals. This view reinforces the argument that the morpho-syntactic kinship of will, can, may and must also reflects semantic kinship. It is suggested that all the modal auxiliaries show that the correspondence relation is non-verified but potential. And the specific place that will holds is that the correspondence is unverified at the time of utterance but will turn out to become verified. The overall conclusion is that idiosyncratic morpho-syntax shared by the modals reflects the semantics and pragmatics of the English modal auxiliaries and is forced also to include will.

  • PDF

A comparison of normalized formant trajectories of English vowels produced by American men and women

  • Yang, Byunggon
    • Phonetics and Speech Sciences
    • /
    • v.11 no.1
    • /
    • pp.1-8
    • /
    • 2019
  • Formant trajectories reflect the continuous variation of speakers' articulatory movements over time. This study examined formant trajectories of English vowels produced by ninety-three American men and women; the values were normalized using the scale function in R and compared using generalized additive mixed models (GAMMs). Praat was used to read the sound data of Hillenbrand et al. (1995). A formant analysis script was prepared, and six formant values at the corresponding time points within each vowel segment were collected. The results indicate that women yielded proportionately higher formant values than men. The standard deviations of each group showed similar patterns at the first formant (F1) and the second formant (F2) axes and at the measurement points. R was used to scale the first two formant data sets of men and women separately. GAMMs of all the scaled formant data produced various patterns of deviation along the measurement points. Generally, more group difference exists in F1 than in F2. Also, women's trajectories appear more dynamic along the vertical and horizontal axes than those of men. The trajectories are related acoustically to F1 and F2 and anatomically to jaw opening and tongue position. We conclude that scaling and nonlinear testing are useful tools for pinpointing differences between speaker group's formant trajectories. This research could be useful as a foundation for future studies comparing curvilinear data sets.

Dynamic Characteristics of Coaxial Swirl-jet Injector with Acoustic Excitation (동축형 스월-제트 분사기의 음향가진에 따른 동특성)

  • Bae, Jinhyun;Kim, Taesung;Jeong, Seokgyu;Jeong, Chanyeong;Choi, Jeong Yeol;Yoon, Youngbin
    • Journal of the Korean Society of Propulsion Engineers
    • /
    • v.22 no.4
    • /
    • pp.99-107
    • /
    • 2018
  • In this study, the injector transfer function (ITF) of a gas-gas coaxial jet-swirl injector is measured by perturbing jet or swirl flow using a speaker as jet flow increases. As a result of measuring the ITF varying feed system length, a peak occurs at a resonance frequency of space where the perturbed flow passes. With jet excitation, the ITF magnitude decreases, but increases thereafter as increasing the jet flow. Therefore the larger the velocity difference between jet and swirl flow, the larger the ITF. With swirl excitation, ITF decreases as increasing the jet flow because of the energy decrease with respect to the constant downstream flow.

A Study on the Development of Embedded Serial Multi-modal Biometrics Recognition System (임베디드 직렬 다중 생체 인식 시스템 개발에 관한 연구)

  • Kim, Joeng-Hoon;Kwon, Soon-Ryang
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.16 no.1
    • /
    • pp.49-54
    • /
    • 2006
  • The recent fingerprint recognition system has unstable factors, such as copy of fingerprint patterns and hacking of fingerprint feature point, which mali cause significant system error. Thus, in this research, we used the fingerprint as the main recognition device and then implemented the multi-biometric recognition system in serial using the speech recognition which has been widely used recently. As a multi-biometric recognition system, once the speech is successfully recognized, the fingerprint recognition process is run. In addition, speaker-dependent DTW(Dynamic Time Warping) algorithm is used among existing speech recognition algorithms (VQ, DTW, HMM, NN) for effective real-time process while KSOM (Kohonen Self-Organizing feature Map) algorithm, which is the artificial intelligence method, is applied for the fingerprint recognition system because of its calculation amount. The experiment of multi-biometric recognition system implemented in this research showed 2 to $7\%$ lower FRR (False Rejection Ratio) than single recognition systems using each fingerprints or voice, but zero FAR (False Acceptance Ratio), which is the most important factor in the recognition system. Moreover, there is almost no difference in the recognition time(average 1.5 seconds) comparing with other existing single biometric recognition systems; therefore, it is proved that the multi-biometric recognition system implemented is more efficient security system than single recognition systems based on various experiments.

A Content-based Video Rate-control Algorithm Interfaced to Human-eye (인간과 결합한 내용기반 동영상 율제어)

  • 황재정;진경식;황치규
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.28 no.3C
    • /
    • pp.307-314
    • /
    • 2003
  • In the general multiple video object coder, more interested objects such as speaker or moving object is consistently coded with higher priority. Since the priority of each object may not be fixed in the whole sequence and be variable on frame basis, it must be adjusted in a frame. In this paper, we analyze the independent rate control algorithm and global algorithm that the QP value is controled by the static parameters, object importance or priority, target PSNR, weighted distortion. The priority among static parameters is analyzed and adjusted into dynamic parameters according to the visual interests or importance obtained by camera interface. Target PSNR and weighted distortion are proportionally derived by using magnitude, motion, and distortion. We apply those parameters for the weighted distortion control and the priority-based control resulting in the efficient bit-rate distribution. As results of this paper, we achieved that fewer bits are allocated for video objects which has less importance and more bits for those which has higher visual importance. The duration of stability in the visual quality is reduced to less than 15 frames of the coded sequence. In the aspect of PSNR, the proposed scheme shows higher quality of more than 2d13 against the conventional schemes. Thus the coding scheme interfaced to human- eye proves an efficient video coder dealing with the multiple number of video objects.

A Study on Lip-reading Enhancement Using Time-domain Filter (시간영역 필터를 이용한 립리딩 성능향상에 관한 연구)

  • 신도성;김진영;최승호
    • The Journal of the Acoustical Society of Korea
    • /
    • v.22 no.5
    • /
    • pp.375-382
    • /
    • 2003
  • Lip-reading technique based on bimodal is to enhance speech recognition rate in noisy environment. It is most important to detect the correct lip-image. But it is hard to estimate stable performance in dynamic environment, because of many factors to deteriorate Lip-reading's performance. There are illumination change, speaker's pronunciation habit, versatility of lips shape and rotation or size change of lips etc. In this paper, we propose the IIR filtering in time-domain for the stable performance. It is very proper to remove the noise of speech, to enhance performance of recognition by digital filtering in time domain. While the lip-reading technique in whole lip image makes data massive, the Principal Component Analysis of pre-process allows to reduce the data quantify by detection of feature without loss of image information. For the observation performance of speech recognition using only image information, we made an experiment on recognition after choosing 22 words in available car service. We used Hidden Markov Model by speech recognition algorithm to compare this words' recognition performance. As a result, while the recognition rate of lip-reading using PCA is 64%, Time-domain filter applied to lip-reading enhances recognition rate of 72.4%.

Automatic Recognition of Pitch Accent Using Distributed Time-Delay Recursive Neural Network (분산 시간지연 회귀신경망을 이용한 피치 악센트 자동 인식)

  • Kim Sung-Suk
    • The Journal of the Acoustical Society of Korea
    • /
    • v.25 no.6
    • /
    • pp.277-281
    • /
    • 2006
  • This paper presents a method for the automatic recognition of pitch accents over syllables. The method that we propose is based on the time-delay recursive neural network (TDRNN). which is a neural network classifier with two different representation of dynamic context: the delayed input nodes allow the representation of an explicit trajectory F0(t) along time. while the recursive nodes provide long-term context information that reflects the characteristics of pitch accentuation in spoken English. We apply the TDRNN to pitch accent recognition in two forms: in the normal TDRNN. all of the prosodic features (pitch. energy, duration) are used as an entire set in a single TDRNN. while in the distributed TDRNN. the network consists of several TDRNNs each taking a single prosodic feature as the input. The final output of the distributed TDRNN is weighted sum of the output of individual TDRNN. We used the Boston Radio News Corpus (BRNC) for the experiments on the speaker-independent pitch accent recognition. π 1e experimental results show that the distributed TDRNN exhibits an average recognition accuracy of 83.64% over both pitch events and non-events.