• Title/Summary/Keyword: Grammar

Search Result 672, Processing Time 0.017 seconds

Korean Sentence Generation Using Phoneme-Level LSTM Language Model (한국어 음소 단위 LSTM 언어모델을 이용한 문장 생성)

  • Ahn, SungMahn;Chung, Yeojin;Lee, Jaejoon;Yang, Jiheon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.71-88
    • /
    • 2017
  • Language models were originally developed for speech recognition and language processing. Using a set of example sentences, a language model predicts the next word or character based on sequential input data. N-gram models have been widely used but this model cannot model the correlation between the input units efficiently since it is a probabilistic model which are based on the frequency of each unit in the training set. Recently, as the deep learning algorithm has been developed, a recurrent neural network (RNN) model and a long short-term memory (LSTM) model have been widely used for the neural language model (Ahn, 2016; Kim et al., 2016; Lee et al., 2016). These models can reflect dependency between the objects that are entered sequentially into the model (Gers and Schmidhuber, 2001; Mikolov et al., 2010; Sundermeyer et al., 2012). In order to learning the neural language model, texts need to be decomposed into words or morphemes. Since, however, a training set of sentences includes a huge number of words or morphemes in general, the size of dictionary is very large and so it increases model complexity. In addition, word-level or morpheme-level models are able to generate vocabularies only which are contained in the training set. Furthermore, with highly morphological languages such as Turkish, Hungarian, Russian, Finnish or Korean, morpheme analyzers have more chance to cause errors in decomposition process (Lankinen et al., 2016). Therefore, this paper proposes a phoneme-level language model for Korean language based on LSTM models. A phoneme such as a vowel or a consonant is the smallest unit that comprises Korean texts. We construct the language model using three or four LSTM layers. Each model was trained using Stochastic Gradient Algorithm and more advanced optimization algorithms such as Adagrad, RMSprop, Adadelta, Adam, Adamax, and Nadam. Simulation study was done with Old Testament texts using a deep learning package Keras based the Theano. After pre-processing the texts, the dataset included 74 of unique characters including vowels, consonants, and punctuation marks. Then we constructed an input vector with 20 consecutive characters and an output with a following 21st character. Finally, total 1,023,411 sets of input-output vectors were included in the dataset and we divided them into training, validation, testsets with proportion 70:15:15. All the simulation were conducted on a system equipped with an Intel Xeon CPU (16 cores) and a NVIDIA GeForce GTX 1080 GPU. We compared the loss function evaluated for the validation set, the perplexity evaluated for the test set, and the time to be taken for training each model. As a result, all the optimization algorithms but the stochastic gradient algorithm showed similar validation loss and perplexity, which are clearly superior to those of the stochastic gradient algorithm. The stochastic gradient algorithm took the longest time to be trained for both 3- and 4-LSTM models. On average, the 4-LSTM layer model took 69% longer training time than the 3-LSTM layer model. However, the validation loss and perplexity were not improved significantly or became even worse for specific conditions. On the other hand, when comparing the automatically generated sentences, the 4-LSTM layer model tended to generate the sentences which are closer to the natural language than the 3-LSTM model. Although there were slight differences in the completeness of the generated sentences between the models, the sentence generation performance was quite satisfactory in any simulation conditions: they generated only legitimate Korean letters and the use of postposition and the conjugation of verbs were almost perfect in the sense of grammar. The results of this study are expected to be widely used for the processing of Korean language in the field of language processing and speech recognition, which are the basis of artificial intelligence systems.

Musical Analysis of Jindo Dasiraegi music for the Scene of Performing Arts Contents (연희현장에서의 올바른 활용을 위한 진도다시래기 음악분석)

  • Han, Seung Seok;Nam, Cho Long
    • (The) Research of the performance art and culture
    • /
    • no.25
    • /
    • pp.253-289
    • /
    • 2012
  • Dasiraegi is a traditional funeral rite performance of Jindo located in the South Jeolla Province of South Korea. With its unique stylistic structure including various dances, songs and witty dialogues, and a storyline depicting the birth of a new life in the wake of death, embodying the Buddhism belief that life and death is interconnected; it attracted great interest from performance organizers and performers who were desperately seeking new contents that can be put on stage as a performance. It is needless to say previous research on Dasiraegi had been most valuable in its recreation as it analyzed the performance from a wide range of perspectives. Despite its contributions, the previous researches were mainly academic focusing on: the symbolic meanings of the performance, basic introduction to the components of the performance such as script, lyrics, witty dialogue, appearance (costume and make-up), stage properties, rhythm, dance and etc., lacking accurate representation of the most crucial element of the performance which is sori (song). For this reason, the study analyzes the music of Dasiraegi and presents its musical characteristics along with its scores to provide practical support for performers who are active in the field. Out of all the numbers in Dasiraegi, this study analyzed all of Geosa-nori and Sadang-nori, the funeral dirge (mourning chant) sung as the performers come on stage and Gasangjae-nori, because among the five proceedings of the funeral rite they were the most commonly performed. There are a plethora of performance recordings to choose from, however, this study chose Jindo Dasiraegi, an album released by E&E Media. The album offers high quality recordings of performances, but more importantly, it is easy to obtain and utilize for performers who want to learn the Dasiraegi based on the script provided in this study. The musical analysis discovered a number of interesting findings. Firstly, most of the songs in Dasiraegi use a typical Yukjabaegi-tori which applies the Mi scale frequently containing cut-off (breaking) sounds. Although, Southern Kyoung-tori which applies the Sol scale was used, it was only in limited parts and was musically incomplete. Secondly, there was no musical affinity between Ssitgim-gut and Dasiraegi albeit both are for funeral rites. The fundamental difference in character and function of Ssitgim-gut and Dasiraegi may be the reason behind this lack of affinity, as Ssitgim-gut is sung to guide the deceased to heaven by comforting him/her, whereas, Dasiaregi is sung to reinvigorate the lives of the living. Lastly, traces of musical grammar found in Pansori are present in the earlier part of Dasiraegi. This may be attributed to the master artist (Designee of Important Intangible Cultural Heritage), who was instrumental in the restoration and hand-down of Dasiaregi, and his experience in a Changgeuk company. The performer's experience with Changgeuk may have induced the alterations in Dasiraegi, causing it to deviate from its original form. On the other hand, it expanded the performative bais by enhancing the performance aspect of Dasiraegi allowing it to be utilized as contents for Performing Arts. It would be meaningful to see this study utilized to benefit future performance artists, taking Dasiraegi as their inspiration, which overcomes the loss of death and invigorates the vibrancy of life.