• Title/Summary/Keyword: LSTM model

Search Result 672, Processing Time 0.03 seconds

LSTM RNN-based Korean Speech Recognition System Using CTC (CTC를 이용한 LSTM RNN 기반 한국어 음성인식 시스템)

  • Lee, Donghyun;Lim, Minkyu;Park, Hosung;Kim, Ji-Hwan
    • Journal of Digital Contents Society
    • /
    • v.18 no.1
    • /
    • pp.93-99
    • /
    • 2017
  • A hybrid approach using Long Short Term Memory (LSTM) Recurrent Neural Network (RNN) has showed great improvement in speech recognition accuracy. For training acoustic model based on hybrid approach, it requires forced alignment of HMM state sequence from Gaussian Mixture Model (GMM)-Hidden Markov Model (HMM). However, high computation time for training GMM-HMM is required. This paper proposes an end-to-end approach for LSTM RNN-based Korean speech recognition to improve learning speed. A Connectionist Temporal Classification (CTC) algorithm is proposed to implement this approach. The proposed method showed almost equal performance in recognition rate, while the learning speed is 1.27 times faster.

An Empirical Study on the Comparison of LSTM and ARIMA Forecasts using Stock Closing Prices

  • Gui Yeol Ryu
    • International journal of advanced smart convergence
    • /
    • v.12 no.1
    • /
    • pp.18-30
    • /
    • 2023
  • We compared empirically the forecast accuracies of the LSTM model, and the ARIMA model. ARIMA model used auto.arima function. Data used in the model is 100 days. We compared with the forecast results for 50 days. We collected the stock closing prices of the top 4 companies by market capitalization in Korea such as "Samsung Electronics", and "LG Energy", "SK Hynix", "Samsung Bio". The collection period is from June 17, 2022, to January 20, 2023. The paired t-test is used to compare the accuracy of forecasts by the two methods because conditions are same. The null hypothesis that the accuracy of the two methods for the four stock closing prices were the same were rejected at the significance level of 5%. Graphs and boxplots confirmed the results of the hypothesis tests. The accuracies of ARIMA are higher than those of LSTM for four cases. For closing stock price of Samsung Electronics, the mean difference of error between ARIMA and LSTM is -370.11, which is 0.618% of the average of the closing stock price. For closing stock price of LG Energy, the mean difference is -4143.298 which is 0.809% of the average of the closing stock price. For closing stock price of SK Hynix, the mean difference is -830.7269 which is 1.00% of the average of the closing stock price. For closing stock price of Samsung Bio, the mean difference is -4143.298 which is 0.809% of the average of the closing stock price. The auto.arima function was used to find the ARIMA model, but other methods are worth considering in future studies. And more efforts are needed to find parameters that provide an optimal model in LSTM.

A Method of Detection of Deepfake Using Bidirectional Convolutional LSTM (Bidirectional Convolutional LSTM을 이용한 Deepfake 탐지 방법)

  • Lee, Dae-hyeon;Moon, Jong-sub
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.30 no.6
    • /
    • pp.1053-1065
    • /
    • 2020
  • With the recent development of hardware performance and artificial intelligence technology, sophisticated fake videos that are difficult to distinguish with the human's eye are increasing. Face synthesis technology using artificial intelligence is called Deepfake, and anyone with a little programming skill and deep learning knowledge can produce sophisticated fake videos using Deepfake. A number of indiscriminate fake videos has been increased significantly, which may lead to problems such as privacy violations, fake news and fraud. Therefore, it is necessary to detect fake video clips that cannot be discriminated by a human eyes. Thus, in this paper, we propose a deep-fake detection model applied with Bidirectional Convolution LSTM and Attention Module. Unlike LSTM, which considers only the forward sequential procedure, the model proposed in this paper uses the reverse order procedure. The Attention Module is used with a Convolutional neural network model to use the characteristics of each frame for extraction. Experiments have shown that the model proposed has 93.5% accuracy and AUC is up to 50% higher than the results of pre-existing studies.

Concrete Crack Prediction Model Combining CNN and LSTM (CNN과 LSTM을 결합한 콘크리트 균열 예측 모델)

  • Dong Eun Lee;Sung Jin Kim;Young Hyun Yoon;Jai Soon Baek
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2023.07a
    • /
    • pp.73-74
    • /
    • 2023
  • 본 논문은 교량 안전에 관련하여 CNN과 LSTM을 결합한 모델을 사용해 콘크리트 균열을 미리 에측한다. 이미지 데이터는 CNN을 통해 처리되고, 시계열 데이러는 LSTM을 통해 처리가 된다. 훈련된 모델을 사용해 새로운 이미지와 시계열 데이터에 대한 균열 예측을 수행한다.

  • PDF

Performance Assessment of Two-stream Convolutional Long- and Short-term Memory Model for September Arctic Sea Ice Prediction from 2001 to 2021 (Two-stream Convolutional Long- and Short-term Memory 모델의 2001-2021년 9월 북극 해빙 예측 성능 평가)

  • Chi, Junhwa
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.6_1
    • /
    • pp.1047-1056
    • /
    • 2022
  • Sea ice, frozen sea water, in the Artic is a primary indicator of global warming. Due to its importance to the climate system, shipping-route navigation, and fisheries, Arctic sea ice prediction has gained increased attention in various disciplines. Recent advances in artificial intelligence (AI), motivated by a desire to develop more autonomous and efficient future predictions, have led to the development of new sea ice prediction models as alternatives to conventional numerical and statistical prediction models. This study aims to evaluate the performance of the two-stream convolutional long-and short-term memory (TS-ConvLSTM) AI model, which is designed for learning both global and local characteristics of the Arctic sea ice changes, for the minimum September Arctic sea ice from 2001 to 2021, and to show the possibility for an operational prediction system. Although the TS-ConvLSTM model generally increased the prediction performance as training data increased, predictability for the marginal ice zone, 5-50% concentration, showed a negative trend due to increasing first-year sea ice and warming. Additionally, a comparison of sea ice extent predicted by the TS-ConvLSTM with the median Sea Ice Outlooks (SIOs) submitted to the Sea Ice Prediction Network has been carried out. Unlike the TS-ConvLSTM, the median SIOs did not show notable improvements as time passed (i.e., the amount of training data increased). Although the TS-ConvLSTM model has shown the potential for the operational sea ice prediction system, learning more spatio-temporal patterns in the difficult-to-predict natural environment for the robust prediction system should be considered in future work.

Korean Sentence Generation Using Phoneme-Level LSTM Language Model (한국어 음소 단위 LSTM 언어모델을 이용한 문장 생성)

  • Ahn, SungMahn;Chung, Yeojin;Lee, Jaejoon;Yang, Jiheon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.71-88
    • /
    • 2017
  • Language models were originally developed for speech recognition and language processing. Using a set of example sentences, a language model predicts the next word or character based on sequential input data. N-gram models have been widely used but this model cannot model the correlation between the input units efficiently since it is a probabilistic model which are based on the frequency of each unit in the training set. Recently, as the deep learning algorithm has been developed, a recurrent neural network (RNN) model and a long short-term memory (LSTM) model have been widely used for the neural language model (Ahn, 2016; Kim et al., 2016; Lee et al., 2016). These models can reflect dependency between the objects that are entered sequentially into the model (Gers and Schmidhuber, 2001; Mikolov et al., 2010; Sundermeyer et al., 2012). In order to learning the neural language model, texts need to be decomposed into words or morphemes. Since, however, a training set of sentences includes a huge number of words or morphemes in general, the size of dictionary is very large and so it increases model complexity. In addition, word-level or morpheme-level models are able to generate vocabularies only which are contained in the training set. Furthermore, with highly morphological languages such as Turkish, Hungarian, Russian, Finnish or Korean, morpheme analyzers have more chance to cause errors in decomposition process (Lankinen et al., 2016). Therefore, this paper proposes a phoneme-level language model for Korean language based on LSTM models. A phoneme such as a vowel or a consonant is the smallest unit that comprises Korean texts. We construct the language model using three or four LSTM layers. Each model was trained using Stochastic Gradient Algorithm and more advanced optimization algorithms such as Adagrad, RMSprop, Adadelta, Adam, Adamax, and Nadam. Simulation study was done with Old Testament texts using a deep learning package Keras based the Theano. After pre-processing the texts, the dataset included 74 of unique characters including vowels, consonants, and punctuation marks. Then we constructed an input vector with 20 consecutive characters and an output with a following 21st character. Finally, total 1,023,411 sets of input-output vectors were included in the dataset and we divided them into training, validation, testsets with proportion 70:15:15. All the simulation were conducted on a system equipped with an Intel Xeon CPU (16 cores) and a NVIDIA GeForce GTX 1080 GPU. We compared the loss function evaluated for the validation set, the perplexity evaluated for the test set, and the time to be taken for training each model. As a result, all the optimization algorithms but the stochastic gradient algorithm showed similar validation loss and perplexity, which are clearly superior to those of the stochastic gradient algorithm. The stochastic gradient algorithm took the longest time to be trained for both 3- and 4-LSTM models. On average, the 4-LSTM layer model took 69% longer training time than the 3-LSTM layer model. However, the validation loss and perplexity were not improved significantly or became even worse for specific conditions. On the other hand, when comparing the automatically generated sentences, the 4-LSTM layer model tended to generate the sentences which are closer to the natural language than the 3-LSTM model. Although there were slight differences in the completeness of the generated sentences between the models, the sentence generation performance was quite satisfactory in any simulation conditions: they generated only legitimate Korean letters and the use of postposition and the conjugation of verbs were almost perfect in the sense of grammar. The results of this study are expected to be widely used for the processing of Korean language in the field of language processing and speech recognition, which are the basis of artificial intelligence systems.

Prediction of Urban Flood Extent by LSTM Model and Logistic Regression (LSTM 모형과 로지스틱 회귀를 통한 도시 침수 범위의 예측)

  • Kim, Hyun Il;Han, Kun Yeun;Lee, Jae Yeong
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.40 no.3
    • /
    • pp.273-283
    • /
    • 2020
  • Because of climate change, the occurrence of localized and heavy rainfall is increasing. It is important to predict floods in urban areas that have suffered inundation in the past. For flood prediction, not only numerical analysis models but also machine learning-based models can be applied. The LSTM (Long Short-Term Memory) neural network used in this study is appropriate for sequence data, but it demands a lot of data. However, rainfall that causes flooding does not appear every year in a single urban basin, meaning it is difficult to collect enough data for deep learning. Therefore, in addition to the rainfall observed in the study area, the observed rainfall in another urban basin was applied in the predictive model. The LSTM neural network was used for predicting the total overflow, and the result of the SWMM (Storm Water Management Model) was applied as target data. The prediction of the inundation map was performed by using logistic regression; the independent variable was the total overflow and the dependent variable was the presence or absence of flooding in each grid. The dependent variable of logistic regression was collected through the simulation results of a two-dimensional flood model. The input data of the two-dimensional flood model were the overflow at each manhole calculated by the SWMM. According to the LSTM neural network parameters, the prediction results of total overflow were compared. Four predictive models were used in this study depending on the parameter of the LSTM. The average RMSE (Root Mean Square Error) for verification and testing was 1.4279 ㎥/s, 1.0079 ㎥/s for the four LSTM models. The minimum RMSE of the verification and testing was calculated as 1.1655 ㎥/s and 0.8797 ㎥/s. It was confirmed that the total overflow can be predicted similarly to the SWMM simulation results. The prediction of inundation extent was performed by linking the logistic regression with the results of the LSTM neural network, and the maximum area fitness was 97.33 % when more than 0.5 m depth was considered. The methodology presented in this study would be helpful in improving urban flood response based on deep learning methodology.

Prediction of Sea Surface Temperature and Detection of Ocean Heat Wave in the South Sea of Korea Using Time-series Deep-learning Approaches (시계열 기계학습을 이용한 한반도 남해 해수면 온도 예측 및 고수온 탐지)

  • Jung, Sihun;Kim, Young Jun;Park, Sumin;Im, Jungho
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.5_3
    • /
    • pp.1077-1093
    • /
    • 2020
  • Sea Surface Temperature (SST) is an important environmental indicator that affects climate coupling systems around the world. In particular, coastal regions suffer from abnormal SST resulting in huge socio-economic damage. This study used Long Short Term Memory (LSTM) and Convolutional Long Short Term Memory (ConvLSTM) to predict SST up to 7 days in the south sea region in South Korea. The results showed that the ConvLSTM model outperformed the LSTM model, resulting in a root mean square error (RMSE) of 0.33℃ and a mean difference of -0.0098℃. Seasonal comparison also showed the superiority of ConvLSTM to LSTM for all seasons. However, in summer, the prediction accuracy for both models with all lead times dramatically decreased, resulting in RMSEs of 0.48℃ and 0.27℃ for LSTM and ConvLSTM, respectively. This study also examined the prediction of abnormally high SST based on three ocean heatwave categories (i.e., warning, caution, and attention) with the lead time from one to seven days for an ocean heatwave case in summer 2017. ConvLSTM was able to successfully predict ocean heatwave five days in advance.

Water Temperature Prediction Study Using Feature Extraction and Reconstruction based on LSTM-Autoencoder

  • Gu-Deuk Song;Su-Hyun Park
    • Journal of the Korea Society of Computer and Information
    • /
    • v.28 no.11
    • /
    • pp.13-20
    • /
    • 2023
  • In this paper, we propose a water temperature prediction method using feature extraction and reconstructed data based on LSTM-Autoencoder. We used multivariate time series data such as sea surface water temperature in the Naksan area of the East Sea where the cold water zone phenomenon occurred, and wind direction and wind speed that affect water temperature. Using the LSTM-Autoencoder model, we used three types of data: feature data extracted through dimensionality reduction of the original data combined with multivariate data of the original data, reconstructed data, and original data. The three types of data were trained by the LSTM model to predict sea surface water temperature and evaluated the accuracy. As a result, the sea surface water temperature prediction accuracy using feature extraction of LSTM-Autoencoder confirmed the best performance with MAE 0.3652, RMSE 0.5604, MAPE 3.309%. The result of this study are expected to be able to prevent damage from natural disasters by improving the prediction accuracy of sea surface temperature changes rapidly such as the cold water zone.

Study on the Prediction of Motion Response of Fishing Vessels using Recurrent Neural Networks (순환 신경망 모델을 이용한 소형어선의 운동응답 예측 연구)

  • Janghoon Seo;Dong-Woo Park;Dong Nam
    • Journal of the Korean Society of Marine Environment & Safety
    • /
    • v.29 no.5
    • /
    • pp.505-511
    • /
    • 2023
  • In the present study, a deep learning model was established to predict the motion response of small fishing vessels. Hydrodynamic performances were evaluated for two small fishing vessels for the dataset of deep learning model. The deep learning model of the Long Short-Term Memory (LSTM) which is one of the recurrent neural network was utilized. The input data of LSTM model consisted of time series of six(6) degrees of freedom motions and wave height and the output label was selected as the time series data of six(6) degrees of freedom motions. The hyperparameter and input window length studies were performed to optimize LSTM model. The time series motion response according to different wave direction was predicted by establised LSTM. The predicted time series motion response showed good overall agreement with the analysis results. As the length of the time series increased, differences between the predicted values and analysis results were increased, which is due to the reduced influence of long-term data in the training process. The overall error of the predicted data indicated that more than 85% of the data showed an error within 10%. The established LSTM model is expected to be utilized in monitoring and alarm systems for small fishing vessels.