• Title/Summary/Keyword: speeches

Search Result 93, Processing Time 0.03 seconds

Gender-Based Differences in Expository Language Use: A Corpus Study of Japanese

  • Heffernan, Kevin;Nishino, Keiko
    • Asia Pacific Journal of Corpus Research
    • /
    • v.1 no.2
    • /
    • pp.1-14
    • /
    • 2020
  • Previous work has shown that men both explain and value the act of explaining more than women, as explaining conveys expertise. However, previous studies are limited to English. We conducted an exploratory study to see if similar patterns are seen amongst Japanese speakers. We examined three registers of Japanese: conversational interviews, simulated speeches, and academic presentations. For each text, we calculated two measures: lexical density and the percentage of the text written in kanji. Both are indicators of expository language. Men produced significantly higher scores for the interviews and speeches. However, the results for the presentations depend on age and academic field. In fields in which women are the minority, women produce higher scores. In the field in which men are the minority, younger men produced higher scores but older men produced lower scores than women of the same age. Our results show that in academic contexts, the explainers are not necessarily men but rather the gender minority. We argue that such speakers are under social pressure to present themselves as experts. These results show that the generalization that men tend to explain more than women does not always hold true, and we urge more academic work on expository language.

Voice Frequency Synthesis using VAW-GAN based Amplitude Scaling for Emotion Transformation

  • Kwon, Hye-Jeong;Kim, Min-Jeong;Baek, Ji-Won;Chung, Kyungyong
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.2
    • /
    • pp.713-725
    • /
    • 2022
  • Mostly, artificial intelligence does not show any definite change in emotions. For this reason, it is hard to demonstrate empathy in communication with humans. If frequency modification is applied to neutral emotions, or if a different emotional frequency is added to them, it is possible to develop artificial intelligence with emotions. This study proposes the emotion conversion using the Generative Adversarial Network (GAN) based voice frequency synthesis. The proposed method extracts a frequency from speech data of twenty-four actors and actresses. In other words, it extracts voice features of their different emotions, preserves linguistic features, and converts emotions only. After that, it generates a frequency in variational auto-encoding Wasserstein generative adversarial network (VAW-GAN) in order to make prosody and preserve linguistic information. That makes it possible to learn speech features in parallel. Finally, it corrects a frequency by employing Amplitude Scaling. With the use of the spectral conversion of logarithmic scale, it is converted into a frequency in consideration of human hearing features. Accordingly, the proposed technique provides the emotion conversion of speeches in order to express emotions in line with artificially generated voices or speeches.

Language style of the President of the Republic of Korea: linguistic and psychological properties in Presidents' Address (대한민국 대통령의 언어스타일: 연설문에 나타난 언어적 특성과 심리적 특성)

  • Kim, Youngjun;Kim, Kyungil
    • Korean Journal of Cognitive Science
    • /
    • v.30 no.3
    • /
    • pp.105-132
    • /
    • 2019
  • This study quantitatively analyzed the speeches of the former 8 Presidents of the Republic of Korea. Unlike the previous studies, this study analyzed both the content word and the functional word related to the psychological phenomenon. In Study 1, six language style indicators such as clout, authenticity, presidentiality, cognitive complexity, femininity, and psychological health were developed. Also, as presidential speeches have been made recently, authenticity, presidentiality, psychological health tended to increase, but cognitive complexity tended to decrease. These tendencies were supported by linear trend analysis. The linguistic differences were distinct not only among individual presidents but also by groups based on the political orientation of the president. In the progressive presidential speech, more clout, cognitive complexity, femininity were observed. In contrast, more authenticity and presidentiality were observed in the conservative presidential speech. In Study 2, it was demonstrated whether the developed language style indicators show reliability in the most recent President's speeches. Finally, it was discussed that the relationship between language style indicators, fitness of the political tendency between the regime and the general public since 2000s, and the characteristics of femininity indicator.

Speech Visualization of Korean Vowels Based on the Distances Among Acoustic Features (음성특징의 거리 개념에 기반한 한국어 모음 음성의 시각화)

  • Pok, Gouchol
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.12 no.5
    • /
    • pp.512-520
    • /
    • 2019
  • It is quite useful to represent speeches visually for learners who study foreign languages as well as the hearing impaired who cannot directly hear speeches, and a number of researches have been presented in the literature. They remain, however, at the level of representing the characteristics of speeches using colors or showing the changing shape of lips and mouth using the animation-based representation. As a result of such approaches, those methods cannot tell the users how far their pronunciations are away from the standard ones, and moreover they make it technically difficult to develop such a system in which users can correct their pronunciation in an interactive manner. In order to address these kind of drawbacks, this paper proposes a speech visualization model based on the relative distance between the user's speech and the standard one, furthermore suggests actual implementation directions by applying the proposed model to the visualization of Korean vowels. The method extract three formants F1, F2, and F3 from speech signals and feed them into the Kohonen's SOM to map the results into 2-D screen and represent each speech as a pint on the screen. We have presented a real system implemented using the open source formant analysis software on the speech of a Korean instructor and several foreign students studying Korean language, in which the user interface was built using the Javascript for the screen display.

Acoustic Channel Compensation at Mel-frequency Spectrum Domain

  • Jeong, So-Young;Oh, Sang-Hoon;Lee, Soo-Young
    • The Journal of the Acoustical Society of Korea
    • /
    • v.22 no.1E
    • /
    • pp.43-48
    • /
    • 2003
  • The effects of linear acoustic channels have been analyzed and compensated at mel-frequency feature domain. Unlike popular RASTA filtering our approach incorporates separate filters for each mel-frequency band, which results in better recognition performance for heavy-reverberated speeches.

Entrepreneur Speech and User Comments: Focusing on YouTube Contents (기업가 연설문의 주제와 시청자 댓글 간의 관계 분석: 유튜브 콘텐츠를 중심으로)

  • Kim, Sungbum;Lee, Junghwan
    • The Journal of the Korea Contents Association
    • /
    • v.20 no.5
    • /
    • pp.513-524
    • /
    • 2020
  • Recently, YouTube's growth started drawing attention. YouTube is not only a content-consumption channel but also provides a space for consumers to express their intention. Consumers share their opinions on YouTube through comments. The study focuses on the text of global entrepreneurs' speeches and the comments in response to those speeches on YouTube. A content analysis was conducted for each speech and comment using the text mining software Leximancer. We analyzed the theme of each entrepreneurial speech and derived topics related to the propensity and characteristics of individual entrepreneurs. In the comments, we found the theme of money, work and need to be common regardless of the content of each speech. Talking into account the different lengths of text, we additionally performed a Prominence Index analysis. We derived time, future, better, best, change, life, business, and need as common keywords for speech contents and viewer comments. Users who watched an entrepreneur's speech on YouTube responded equally to the topics of life, time, future, customer needs, and positive change.

A Study on the Multilingual Speech Recognition for On-line International Game (온라인 다국적 게임을 위한 다국어 혼합 음성 인식에 관한 연구)

  • Kim, Suk-Dong;Kang, Heung-Soon;Woo, In-Sung;Shin, Chwa-Cheul;Yoon, Chun-Duk
    • Journal of Korea Game Society
    • /
    • v.8 no.4
    • /
    • pp.107-114
    • /
    • 2008
  • The requests for speech-recognition for multi-language in field of game and the necessity of multi-language system, which expresses one phonetic model from many different kind of language phonetics, has been increased in field of game industry. Here upon, the research regarding development of multi-national language system which can express speeches, that is consist of various different languages, into only one lexical model is needed. In this paper is basic research for establishing integrated system from multi-language lexical model, and it shows the system which recognize Korean and English speeches into IPA(International Phonetic Alphabet). We focused on finding the IPA model which is satisfied with Korean and English phoneme one simutaneously. As a result, we could get the 90.62% of Korean speech-recognition rate, also 91.71% of English speech-recognition rate.

  • PDF

Acoustic Masking Effect That Can Be Occurred by Speech Contrast Enhancement in Hearing Aids (보청기에서 음성 대비 강조에 의해 발생할 수 있는 마스킹 현상)

  • Jeon, Y.Y.;Yang, D.G.;Bang, D.H.;Kil, S.K.;Lee, S.M.
    • Journal of rehabilitation welfare engineering & assistive technology
    • /
    • v.1 no.1
    • /
    • pp.21-28
    • /
    • 2007
  • In most of hearing aids, amplification algorithms are used to compensate hearing loss, noise and feedback reduction algorithms are used and to increase the perception of speeches contrast enhancement algorithms are used. However, acoustic masking effect is occurred between formants if contrast is enhanced excessively. To confirm the masking effect in speeches, the experiment are composed of 6 tests; test pure tone test, speech reception test, word recognition test, pure tone masking test, formant pure tone masking test and speech masking test, and for objective evaluation, LLR is introduced. As a result of normal hearing subjects and hearing impaired subjects, more making is occurred in hearing impaired subjects than normal hearing subjects when using pure tone, and in the speech masking test, speech reception is also lower in hearing impaired subjects than in normal hearing subjects. This means that acoustic masking effect rather than distortion influences speech perception. So it is required to check the characteristics of masking effect before wearing a hearing aid and to apply this characteristics to fitting curve.

  • PDF

Japan's Economic Performance "Lost Decade": Myth, Reality, or Role Model?

  • Anderson, Richard G.
    • The Journal of Asian Finance, Economics and Business
    • /
    • v.3 no.1
    • /
    • pp.15-32
    • /
    • 2016
  • Between 2010 and 2012, former Bank of Japan Governor Shirakawa argued, in series of speeches, that Japan's economic performance, when compared to that of other G7 nations, was stronger after 1990 than appreciated by the critics. In May 2012, Nobel prize-winning economist and Princeton University professor Paul Krugman echoed a similar sentiment in a Financial Times interview. This analysis expands on these assertions and asks to what extent they are supported by cross-section data for the G7. As reviewed below, to date, no idiosyncratic explanation has arisen to explain the Japanese slowdown-perhaps this is the correct explanation: the slowdown in Japan, once adjusted for demographics, is less severe than in other G7 countries and, as recently noted by Eichengreen, Park, and Shin (2015), TFP growth in a number of other nations followed similar patterns. Focused on labor productivity, far from a laggard, Japan's performance ranks near the best in the G7.

Study about Windows System Control Using Gesture and Speech Recognition (제스처 및 음성 인식을 이용한 윈도우 시스템 제어에 관한 연구)

  • 김주홍;진성일이남호이용범
    • Proceedings of the IEEK Conference
    • /
    • 1998.10a
    • /
    • pp.1289-1292
    • /
    • 1998
  • HCI(human computer interface) technologies have been often implemented using mouse, keyboard and joystick. Because mouse and keyboard are used only in limited situation, More natural HCI methods such as speech based method and gesture based method recently attract wide attention. In this paper, we present multi-modal input system to control Windows system for practical use of multi-media computer. Our multi-modal input system consists of three parts. First one is virtual-hand mouse part. This part is to replace mouse control with a set of gestures. Second one is Windows control system using speech recognition. Third one is Windows control system using gesture recognition. We introduce neural network and HMM methods to recognize speeches and gestures. The results of three parts interface directly to CPU and through Windows.

  • PDF