• Title/Summary/Keyword: long-short-term memory

Search Result 623, Processing Time 0.031 seconds

Classification of Behavior of UTD Data using LSTM Technique (LSTM 기법을 적용한 UTD 데이터 행동 분류)

  • Jeung, Gyeo-wun;Ahn, Ji-min;Shin, Dong-in;Won, Geon;Park, Jong-bum
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2018.10a
    • /
    • pp.477-479
    • /
    • 2018
  • This study was carried out to utilize LSTM(Long Short-Term Memory) technique which is one kind of artificial neural network. Among the 27 types of motion data released by the UTD(University of Texas at Dallas), 3-axis acceleration and angular velocity data were applied to the basic LSTM and Deep Residual Bidir-LSTM techniques to classify the behavior.

  • PDF

Integrated Dialogue Analysis using Long Short-Term Memory (Long Short-Term Memory를 이용한 통합 대화 분석)

  • Kim, Min-Kyoung;Kim, Harksoo
    • Annual Conference on Human and Language Technology
    • /
    • 2016.10a
    • /
    • pp.119-121
    • /
    • 2016
  • 최근 사람과 컴퓨터가 대화를 하는 채팅시스템 연구가 활발해지고 있다. 컴퓨터가 사람의 말에 적절한 응답을 하기 위해선 그 의미를 분석할 필요가 있다. 발화에 대한 의미 분석의 기본이 되는 연구로 감정분석과 화행분석이 있다. 그러나 이 둘은 서로 밀접한 연관이 있음에도 불구하고 함께 분석하는 연구가 시도 되지 않았다. 본 연구에서는 Long Short-term Memory(LSTM)를 이용하여 대화체 문장의 감정과 화행, 서술자를 동시에 분석하는 통합 대화 분석모델을 제안한다. 사랑 도메인 데이터를 사용한 실험에서 제안 모델은 감정 58.08%, 화행 82.60%, 서술자 62.74%의 정확도(Accuracy)를 보였다.

  • PDF

Multi-layered attentional peephole convolutional LSTM for abstractive text summarization

  • Rahman, Md. Motiur;Siddiqui, Fazlul Hasan
    • ETRI Journal
    • /
    • v.43 no.2
    • /
    • pp.288-298
    • /
    • 2021
  • Abstractive text summarization is a process of making a summary of a given text by paraphrasing the facts of the text while keeping the meaning intact. The manmade summary generation process is laborious and time-consuming. We present here a summary generation model that is based on multilayered attentional peephole convolutional long short-term memory (MAPCoL; LSTM) in order to extract abstractive summaries of large text in an automated manner. We added the concept of attention in a peephole convolutional LSTM to improve the overall quality of a summary by giving weights to important parts of the source text during training. We evaluated the performance with regard to semantic coherence of our MAPCoL model over a popular dataset named CNN/Daily Mail, and found that MAPCoL outperformed other traditional LSTM-based models. We found improvements in the performance of MAPCoL in different internal settings when compared to state-of-the-art models of abstractive text summarization.

Forecasting Container Throughput with Long Short Term Memory (LSTM을 활용한 컨테이너 물동량 예측)

  • Lim, Sangseop
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2020.07a
    • /
    • pp.617-618
    • /
    • 2020
  • 우리나라의 지리적인 여건상 대륙과 연결되지 않기 때문에 해상운송에 절대적으로 의존하고 있다. 해상운송에 있어 항만시설의 확보가 필요하며 대외무역의존도가 높은 우리나라의 경우 더욱 중요한 역할을 한다. 항만시설은 장기적인 항만수요예측을 통해 대규모 인프라투자를 결정하며 단기적인 예측은 항만운영의 효율성을 개선하고 항만의 경쟁력을 제고하는데 기여하므로 예측의 정확성을 높이기 위해 많은 노력이 필요하다. 본 논문에서는 딥러닝 모델 중에 하나인 LSTM(Long Short Term Memory)을 적용하여 우리나라 주요항만의 컨테이너 물동량 단기예측을 수행하여 선행연구들에서 주류를 이뤘던 ARIMA류의 시계열모델과 비교하여 예측성능을 평가할 것이다. 본 논문은 학문적으로 항만수요예측에 관한 새로운 예측모델을 제시하였다는 측면에서 의미가 있으며 실무적으로 항만수요예측에 대한 정확성을 개선하여 항만투자의사결정에 과학적인 근거로서 활용이 가능할 것으로 기대된다.

  • PDF

A Delta- and Attention-based Long Short-Term Memory (LSTM) Architecture model for Rainfall-runoff Modeling

  • Ahn, Kuk-Hyun;Yoon, Sunghyun
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2022.05a
    • /
    • pp.35-35
    • /
    • 2022
  • 최근에 딥 러닝(Deep learning) 기반의 많은 방법들이 수문학적 모형 및 예측에서 의미있는 결과를 보여주고 있지만 더 많은 연구가 요구되고 있다. 본 연구에서는 수자원의 가장 대표적인 모델링 구조인 강우유출의 관계의 규명에 대한 모형을 Long Short-Term Memory (LSTM) 기반의 변형 된 방법으로 제시하고자 한다. 구체적으로 본 연구에서는 반응변수인 유출량에 대한 직접적인 고려가 아니라 그의 1차 도함수 (First derivative)로 정의되는 Delta기반으로 모형을 구축하였다. 또한, Attention 메카니즘 기반의 모형을 사용함으로써 강우유출의 관계의 규명에 있어 정확성을 향상시키고자 하였다. 마지막으로 확률 기반의 예측를 생성하고 이에 대한 불확실성의 고려를 위하여 Denisty 기반의 모형을 포함시켰고 이를 통하여 Epistemic uncertainty와 Aleatory uncertainty에 대한 상대적 정량화를 수행하였다. 본 연구에서 제시되는 모형의 효용성 및 적용성을 평가하기 위하여 미국 전역에 위치하는 총 507개의 유역의 일별 데이터를 기반으로 모형을 평가하였다. 결과적으로 본 연구에서 제시한 모형이 기존의 대표적인 딥 러닝 기반의 모형인 LSTM 모형과 비교하였을 때 높은 정확성뿐만 아니라 불확실성의 표현과 정량화에 대한 유용한 것으로 확인되었다.

  • PDF

Text Classification Method Using Deep Learning Model Fusion and Its Application

  • Shin, Seong-Yoon;Cho, Gwang-Hyun;Cho, Seung-Pyo;Lee, Hyun-Chang
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.409-410
    • /
    • 2022
  • This paper proposes a fusion model based on Long-Short Term Memory networks (LSTM) and CNN deep learning methods, and applied to multi-category news datasets, and achieved good results. Experiments show that the fusion model based on deep learning has greatly improved the precision and accuracy of text sentiment classification. This method will become an important way to optimize the model and improve the performance of the model.

  • PDF

Prediction of Baltic Dry Index by Applications of Long Short-Term Memory (Long Short-Term Memory를 활용한 건화물운임지수 예측)

  • HAN, Minsoo;YU, Song-Jin
    • Journal of Korean Society for Quality Management
    • /
    • v.47 no.3
    • /
    • pp.497-508
    • /
    • 2019
  • Purpose: The purpose of this study is to overcome limitations of conventional studies that to predict Baltic Dry Index (BDI). The study proposed applications of Artificial Neural Network (ANN) named Long Short-Term Memory (LSTM) to predict BDI. Methods: The BDI time-series prediction was carried out through eight variables related to the dry bulk market. The prediction was conducted in two steps. First, identifying the goodness of fitness for the BDI time-series of specific ANN models and determining the network structures to be used in the next step. While using ANN's generalization capability, the structures determined in the previous steps were used in the empirical prediction step, and the sliding-window method was applied to make a daily (one-day ahead) prediction. Results: At the empirical prediction step, it was possible to predict variable y(BDI time series) at point of time t by 8 variables (related to the dry bulk market) of x at point of time (t-1). LSTM, known to be good at learning over a long period of time, showed the best performance with higher predictive accuracy compared to Multi-Layer Perceptron (MLP) and Recurrent Neural Network (RNN). Conclusion: Applying this study to real business would require long-term predictions by applying more detailed forecasting techniques. I hope that the research can provide a point of reference in the dry bulk market, and furthermore in the decision-making and investment in the future of the shipping business as a whole.

Long-term prediction of safety parameters with uncertainty estimation in emergency situations at nuclear power plants

  • Hyojin Kim;Jonghyun Kim
    • Nuclear Engineering and Technology
    • /
    • v.55 no.5
    • /
    • pp.1630-1643
    • /
    • 2023
  • The correct situation awareness (SA) of operators is important for managing nuclear power plants (NPPs), particularly in accident-related situations. Among the three levels of SA suggested by Ensley, Level 3 SA (i.e., projection of the future status of the situation) is challenging because of the complexity of NPPs as well as the uncertainty of accidents. Hence, several prediction methods using artificial intelligence techniques have been proposed to assist operators in accident prediction. However, these methods only predict short-term plant status (e.g., the status after a few minutes) and do not provide information regarding the uncertainty associated with the prediction. This paper proposes an algorithm that can predict the multivariate and long-term behavior of plant parameters for 2 h with 120 steps and provide the uncertainty of the prediction. The algorithm applies bidirectional long short-term memory and an attention mechanism, which enable the algorithm to predict the precise long-term trends of the parameters with high prediction accuracy. A conditional variational autoencoder was used to provide uncertainty information about the network prediction. The algorithm was trained, optimized, and validated using a compact nuclear simulator for a Westinghouse 900 MWe NPP.

Integrate-and-Fire Neuron Circuit and Synaptic Device with Floating Body MOSFETs

  • Kwon, Min-Woo;Kim, Hyungjin;Park, Jungjin;Park, Byung-Gook
    • JSTS:Journal of Semiconductor Technology and Science
    • /
    • v.14 no.6
    • /
    • pp.755-759
    • /
    • 2014
  • We propose an integrate-and-fire neuron circuit and synaptic devices with the floating body MOSFETs. The synaptic devices consist of a floating body MOSFET to imitate biological synaptic characteristics. The synaptic learning is performed by hole accumulation. The synaptic device has short-term and long-term memory in a single silicon device. I&F neuron circuit emulate the biological neuron characteristics such as integration, threshold triggering, output generation, and refractory period, using floating body MOSFET. The neuron circuit sends feedback signal to the synaptic transistor for long-term memory.

Speaker verification system combining attention-long short term memory based speaker embedding and I-vector in far-field and noisy environments (Attention-long short term memory 기반의 화자 임베딩과 I-vector를 결합한 원거리 및 잡음 환경에서의 화자 검증 알고리즘)

  • Bae, Ara;Kim, Wooil
    • The Journal of the Acoustical Society of Korea
    • /
    • v.39 no.2
    • /
    • pp.137-142
    • /
    • 2020
  • Many studies based on I-vector have been conducted in a variety of environments, from text-dependent short-utterance to text-independent long-utterance. In this paper, we propose a speaker verification system employing a combination of I-vector with Probabilistic Linear Discriminant Analysis (PLDA) and speaker embedding of Long Short Term Memory (LSTM) with attention mechanism in far-field and noisy environments. The LSTM model's Equal Error Rate (EER) is 15.52 % and the Attention-LSTM model is 8.46 %, improving by 7.06 %. We show that the proposed method solves the problem of the existing extraction process which defines embedding as a heuristic. The EER of the I-vector/PLDA without combining is 6.18 % that shows the best performance. And combined with attention-LSTM based embedding is 2.57 % that is 3.61 % less than the baseline system, and which improves performance by 58.41 %.