• 제목/요약/키워드: LSTM model

Search Result 657, Processing Time 0.152 seconds

A Comparative study on smoothing techniques for performance improvement of LSTM learning model

  • Tae-Jin, Park;Gab-Sig, Sim
    • Journal of the Korea Society of Computer and Information
    • /
    • v.28 no.1
    • /
    • pp.17-26
    • /
    • 2023
  • In this paper, we propose a several smoothing techniques are compared and applied to increase the application of the LSTM-based learning model and its effectiveness. The applied smoothing technique is Savitky-Golay, exponential smoothing, and weighted moving average. Through this study, the LSTM algorithm with the Savitky-Golay filter applied in the preprocessing process showed significant best results in prediction performance than the result value shown when applying the LSTM model to Bitcoin data. To confirm the predictive performance results, the learning loss rate and verification loss rate according to the Savitzky-Golay LSTM model were compared with the case of LSTM used to remove complex factors from Bitcoin price prediction, and experimented with an average value of 20 times to increase its reliability. As a result, values of (3.0556, 0.00005) and (1.4659, 0.00002) could be obtained. As a result, since crypto-currencies such as Bitcoin have more volatility than stocks, noise was removed by applying the Savitzky-Golay in the data preprocessing process, and the data after preprocessing were obtained the most-significant to increase the Bitcoin prediction rate through LSTM neural network learning.

LSTM Language Model Based Korean Sentence Generation (LSTM 언어모델 기반 한국어 문장 생성)

  • Kim, Yang-hoon;Hwang, Yong-keun;Kang, Tae-gwan;Jung, Kyo-min
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.41 no.5
    • /
    • pp.592-601
    • /
    • 2016
  • The recurrent neural network (RNN) is a deep learning model which is suitable to sequential or length-variable data. The Long Short-Term Memory (LSTM) mitigates the vanishing gradient problem of RNNs so that LSTM can maintain the long-term dependency among the constituents of the given input sequence. In this paper, we propose a LSTM based language model which can predict following words of a given incomplete sentence to generate a complete sentence. To evaluate our method, we trained our model using multiple Korean corpora then generated the incomplete part of Korean sentences. The result shows that our language model was able to generate the fluent Korean sentences. We also show that the word based model generated better sentences compared to the other settings.

Prediction of time-series underwater noise data using long short term memory model (Long short term memory 모델을 이용한 시계열 수중 소음 데이터 예측)

  • Hyesun Lee;Wooyoung Hong;Kookhyun Kim;Keunhwa Lee
    • The Journal of the Acoustical Society of Korea
    • /
    • v.42 no.4
    • /
    • pp.313-319
    • /
    • 2023
  • In this paper, a time series machine learning model, Long Short Term Memory (LSTM), is applied into the bubble flow noise data and the underwater projectile launch noise data to predict missing values of time-series underwater noise data. The former is mixed with bubble noise, flow noise, and fluid-induced interaction noise measured in a pipe and can be classified into three types. The latter is the noise generated when an underwater projectile is ejected from a launch tube and has a characteristic of instantaenous noise. For such types of noise, a data-driven model can be more useful than an analytical model. We constructed an LSTM model with given data and evaluated the model's performance based on the number of hidden units, the number of input sequences, and the decimation factor of signal. It is shown that the optimal LSTM model works well for new data of the same type.

Fashion Clothing Image Classification Deep Learning (패션 의류 영상 분류 딥러닝)

  • Shin, Seong-Yoon;Wang, Guangxing;Shin, Kwang-Seong;Lee, Hyun-Chang
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.676-677
    • /
    • 2022
  • In this paper, we propose a new method based on a deep learning model with an optimized dynamic decay learning rate and improved model structure to achieve fast and accurate classification of fashion clothing images. Experiments are performed using the model proposed in the Fashion-MNIST dataset and compared with methods of CNN, LeNet, LSTM and BiLSTM.

  • PDF

Research on Chinese Microblog Sentiment Classification Based on TextCNN-BiLSTM Model

  • Haiqin Tang;Ruirui Zhang
    • Journal of Information Processing Systems
    • /
    • v.19 no.6
    • /
    • pp.842-857
    • /
    • 2023
  • Currently, most sentiment classification models on microblogging platforms analyze sentence parts of speech and emoticons without comprehending users' emotional inclinations and grasping moral nuances. This study proposes a hybrid sentiment analysis model. Given the distinct nature of microblog comments, the model employs a combined stop-word list and word2vec for word vectorization. To mitigate local information loss, the TextCNN model, devoid of pooling layers, is employed for local feature extraction, while BiLSTM is utilized for contextual feature extraction in deep learning. Subsequently, microblog comment sentiments are categorized using a classification layer. Given the binary classification task at the output layer and the numerous hidden layers within BiLSTM, the Tanh activation function is adopted in this model. Experimental findings demonstrate that the enhanced TextCNN-BiLSTM model attains a precision of 94.75%. This represents a 1.21%, 1.25%, and 1.25% enhancement in precision, recall, and F1 values, respectively, in comparison to the individual deep learning models TextCNN. Furthermore, it outperforms BiLSTM by 0.78%, 0.9%, and 0.9% in precision, recall, and F1 values.

Improvement of Track Tracking Performance Using Deep Learning-based LSTM Model (딥러닝 기반 LSTM 모형을 이용한 항적 추적성능 향상에 관한 연구)

  • Hwang, Jin-Ha;Lee, Jong-Min
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.189-192
    • /
    • 2021
  • This study applies a deep learning-based long short-term memory(LSTM) model to track tracking technology. In the case of existing track tracking technology, the weight of constant velocity, constant acceleration, stiff turn, and circular(3D) flight is automatically changed when tracking track in real time using LMIPDA based on Kalman filter according to flight characteristics of an aircraft such as constant velocity, constant acceleration, stiff turn, and circular(3D) flight. In this process, it is necessary to improve performance of changing flight characteristic weight, because changing flight characteristics such as stiff turn flight during constant velocity flight could incur the loss of track and decreasing of the tracking performance. This study is for improving track tracking performance by predicting the change of flight characteristics in advance and changing flight characteristic weigh rapidly. To get this result, this study makes deep learning-based Long Short-Term Memory(LSTM) model study the plot and target of simulator applied with radar error model, and compares the flight tracking results of using Kalman filter with those of deep learning-based Long Short-Term memory(LSTM) model.

  • PDF

Emotion Analysis Using a Bidirectional LSTM for Word Sense Disambiguation (양방향 LSTM을 적용한 단어의미 중의성 해소 감정분석)

  • Ki, Ho-Yeon;Shin, Kyung-shik
    • The Journal of Bigdata
    • /
    • v.5 no.1
    • /
    • pp.197-208
    • /
    • 2020
  • Lexical ambiguity means that a word can be interpreted as two or more meanings, such as homonym and polysemy, and there are many cases of word sense ambiguation in words expressing emotions. In terms of projecting human psychology, these words convey specific and rich contexts, resulting in lexical ambiguity. In this study, we propose an emotional classification model that disambiguate word sense using bidirectional LSTM. It is based on the assumption that if the information of the surrounding context is fully reflected, the problem of lexical ambiguity can be solved and the emotions that the sentence wants to express can be expressed as one. Bidirectional LSTM is an algorithm that is frequently used in the field of natural language processing research requiring contextual information and is also intended to be used in this study to learn context. GloVe embedding is used as the embedding layer of this research model, and the performance of this model was verified compared to the model applied with LSTM and RNN algorithms. Such a framework could contribute to various fields, including marketing, which could connect the emotions of SNS users to their desire for consumption.

Prediction of water level in a tidal river using a deep-learning based LSTM model (딥러닝 기반 LSTM 모형을 이용한 감조하천 수위 예측)

  • Jung, Sungho;Cho, Hyoseob;Kim, Jeongyup;Lee, Giha
    • Journal of Korea Water Resources Association
    • /
    • v.51 no.12
    • /
    • pp.1207-1216
    • /
    • 2018
  • Discharge or water level predictions at tidally affected river reaches are currently still a great challenge in hydrological practices. This research aims to predict water level of the tide dominated site, Jamsu bridge in the Han River downstream. Physics-based hydrodynamic approaches are sometimes not applicable for water level prediction in such a tidal river due to uncertainty sources like rainfall forecasting data. In this study, TensorFlow deep learning framework was used to build a deep neural network based LSTM model and its applications. The LSTM model was trained based on 3 data sets having 10-min temporal resolution: Paldang dam release, Jamsu bridge water level, predicted tidal level for 6 years (2011~2016) and then predict the water level time series given the six lead times: 1, 3, 6, 9, 12, 24 hours. The optimal hyper-parameters of LSTM model were set up as follows: 6 hidden layers number, 0.01 learning rate, 3000 iterations. In addition, we changed the key parameter of LSTM model, sequence length, ranging from 1 to 6 hours to test its affect to prediction results. The LSTM model with the 1 hr sequence length led to the best performing prediction results for the all cases. In particular, it resulted in very accurate prediction: RMSE (0.065 cm) and NSE (0.99) for the 1 hr lead time prediction case. However, as the lead time became longer, the RMSE increased from 0.08 m (1 hr lead time) to 0.28 m (24 hrs lead time) and the NSE decreased from 0.99 (1 hr lead time) to 0.74 (24 hrs lead time), respectively.

Development of a Prediction Model of Solar Irradiances Using LSTM for Use in Building Predictive Control (건물 예측 제어용 LSTM 기반 일사 예측 모델)

  • Jeon, Byung-Ki;Lee, Kyung-Ho;Kim, Eui-Jong
    • Journal of the Korean Solar Energy Society
    • /
    • v.39 no.5
    • /
    • pp.41-52
    • /
    • 2019
  • The purpose of the work is to develop a simple solar irradiance prediction model using a deep learning method, the LSTM (long term short term memory). Other than existing prediction models, the proposed one uses only the cloudiness among the information forecasted from the national meterological forecast center. The future cloudiness is generally announced with four categories and for three-hour intervals. In this work, a daily irradiance pattern is used as an input vector to the LSTM together with that cloudiness information. The proposed model showed an error of 5% for learning and 30% for prediction. This level of error has lower influence on the load prediction in typical building cases.

Classification in Different Genera by Cytochrome Oxidase Subunit I Gene Using CNN-LSTM Hybrid Model

  • Meijing Li;Dongkeun Kim
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.2
    • /
    • pp.159-166
    • /
    • 2023
  • The COI gene is a sequence of approximately 650 bp at the 5' terminal of the mitochondrial Cytochrome c Oxidase subunit I (COI) gene. As an effective DeoxyriboNucleic Acid (DNA) barcode, it is widely used for the taxonomic identification and evolutionary analysis of species. We created a CNN-LSTM hybrid model by combining the gene features partially extracted by the Long Short-Term Memory ( LSTM ) network with the feature maps obtained by the CNN. Compared to K-Means Clustering, Support Vector Machines (SVM), and a single CNN classification model, after training 278 samples in a training set that included 15 genera from two orders, the CNN-LSTM hybrid model achieved 94% accuracy in the test set, which contained 118 samples. We augmented the training set samples and four genera into four orders, and the classification accuracy of the test set reached 100%. This study also proposes calculating the cosine similarity between the training and test sets to initially assess the reliability of the predicted results and discover new species.