• 제목/요약/키워드: Recurrent Neural Network

검색결과 566건 처리시간 0.038초

Recurrent Neural Network Adaptive Equalizers Based on Data Communication

  • Jiang, Hongrui;Kwak, Kyung-Sup
    • Journal of Communications and Networks
    • /
    • 제5권1호
    • /
    • pp.7-18
    • /
    • 2003
  • In this paper, a decision feedback recurrent neural network equalizer and a modified real time recurrent learning algorithm are proposed, and an adaptive adjusting of the learning step is also brought forward. Then, a complex case is considered. A decision feedback complex recurrent neural network equalizer and a modified complex real time recurrent learning algorithm are proposed. Moreover, weights of decision feedback recurrent neural network equalizer under burst-interference conditions are analyzed, and two anti-burst-interference algorithms to prevent equalizer from out of working are presented, which are applied to both real and complex cases. The performance of the recurrent neural network equalizer is analyzed based on numerical results.

A New Recurrent Neural Network Architecture for Pattern Recognition and Its Convergence Results

  • Lee, Seong-Whan;Kim, Young-Joon;Song, Hee-Heon
    • Journal of Electrical Engineering and information Science
    • /
    • 제1권1호
    • /
    • pp.108-117
    • /
    • 1996
  • In this paper, we propose a new type of recurrent neural network architecture in which each output unit is connected with itself and fully-connected with other output units and all hidden units. The proposed recurrent network differs from Jordan's and Elman's recurrent networks in view of functions and architectures because it was originally extended from the multilayer feedforward neural network for improving the discrimination and generalization power. We also prove the convergence property of learning algorithm of the proposed recurrent neural network and analyze the performance of the proposed recurrent neural network by performing recognition experiments with the totally unconstrained handwritten numeral database of Concordia University of Canada. Experimental results confirmed that the proposed recurrent neural network improves the discrimination and generalization power in recognizing spatial patterns.

  • PDF

순환신경망 기초 실습 사례 개발 (Development of Basic Practice Cases for Recurrent Neural Networks)

  • 허경
    • 실천공학교육논문지
    • /
    • 제14권3호
    • /
    • pp.491-498
    • /
    • 2022
  • 본 논문에서는 비전공자들을 위한 교양과정으로, 기초 순환신경망 과목 커리큘럼을 설계하는데 필수적으로 요구되는 순환신경망 SW 실습 사례를 개발하였다. 개발된 SW 실습 사례는 순환신경망의 동작원리를 이해시키는 데 초점을 두고, 시각화된 전체 동작 과정을 확인할 수 있도록 스프레드시트를 사용하였다. 개발된 순환신경망 실습 사례는 지도학습 방식의 텍스트완성 훈련데이터 생성, 입력층, 은닉층, 상태층(컨텍스트 노드) 그리고 출력층을 차례대로 구현하고, 텍스트 데이터에 대해 순환신경망의 성능을 테스트하는 것으로 구성되었다. 본 논문에서 개발한 순환신경망 실습사례는 다양한 문자 수를 갖는 단어를 자동 완성한다. 제안한 순환신경망 실습사례를 활용하여, 한글 또는 영어 단어를 구성하는 최대 문자 수를 다양하게 확장하여 자동 완성하는 인공지능 SW 실습 사례를 만들 수 있다. 따라서, 본 순환신경망 기초 실습 사례의 활용도가 높다고 할 수 있다.

Recurrent Neural Network를 이용한 이미지 캡션 생성 (Image Caption Generation using Recurrent Neural Network)

  • 이창기
    • 정보과학회 논문지
    • /
    • 제43권8호
    • /
    • pp.878-882
    • /
    • 2016
  • 이미지의 내용을 설명하는 캡션을 자동으로 생성하는 기술은 이미지 인식과 자연어처리 기술을 필요로 하는 매우 어려운 기술이지만, 유아 교육이나 이미지 검색, 맹인들을 위한 네비게이션 등에 사용될 수 있는 중요한 기술이다. 본 논문에서는 이미지 캡션 생성을 위해 Convolutional Neural Network(CNN)으로 인코딩된 이미지 정보를 입력으로 갖는 이미지 캡션 생성에 최적화된 Recurrent Neural Network(RNN) 모델을 제안하고, 실험을 통해 본 논문에서 제안한 모델이 Flickr 8K와 Flickr 30K, MS COCO 데이터 셋에서 기존의 연구들보다 높은 성능을 얻음을 보인다.

시계열 자료의 예측을 위한 베이지안 순환 신경망에 관한 연구 (A Study on the Bayesian Recurrent Neural Network for Time Series Prediction)

  • 홍찬영;박정훈;윤태성;박진배
    • 제어로봇시스템학회논문지
    • /
    • 제10권12호
    • /
    • pp.1295-1304
    • /
    • 2004
  • In this paper, the Bayesian recurrent neural network is proposed to predict time series data. A neural network predictor requests proper learning strategy to adjust the network weights, and one needs to prepare for non-linear and non-stationary evolution of network weights. The Bayesian neural network in this paper estimates not the single set of weights but the probability distributions of weights. In other words, the weights vector is set as a state vector of state space method, and its probability distributions are estimated in accordance with the particle filtering process. This approach makes it possible to obtain more exact estimation of the weights. In the aspect of network architecture, it is known that the recurrent feedback structure is superior to the feedforward structure for the problem of time series prediction. Therefore, the recurrent neural network with Bayesian inference, what we call Bayesian recurrent neural network (BRNN), is expected to show higher performance than the normal neural network. To verify the proposed method, the time series data are numerically generated and various kinds of neural network predictor are applied on it in order to be compared. As a result, feedback structure and Bayesian learning are better than feedforward structure and backpropagation learning, respectively. Consequently, it is verified that the Bayesian reccurent neural network shows better a prediction result than the common Bayesian neural network.

Bayesian Neural Network with Recurrent Architecture for Time Series Prediction

  • Hong, Chan-Young;Park, Jung-Hun;Yoon, Tae-Sung;Park, Jin-Bae
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2004년도 ICCAS
    • /
    • pp.631-634
    • /
    • 2004
  • In this paper, the Bayesian recurrent neural network (BRNN) is proposed to predict time series data. Among the various traditional prediction methodologies, a neural network method is considered to be more effective in case of non-linear and non-stationary time series data. A neural network predictor requests proper learning strategy to adjust the network weights, and one need to prepare for non-linear and non-stationary evolution of network weights. The Bayesian neural network in this paper estimates not the single set of weights but the probability distributions of weights. In other words, we sets the weight vector as a state vector of state space method, and estimates its probability distributions in accordance with the Bayesian inference. This approach makes it possible to obtain more exact estimation of the weights. Moreover, in the aspect of network architecture, it is known that the recurrent feedback structure is superior to the feedforward structure for the problem of time series prediction. Therefore, the recurrent network with Bayesian inference, what we call BRNN, is expected to show higher performance than the normal neural network. To verify the performance of the proposed method, the time series data are numerically generated and a neural network predictor is applied on it. As a result, BRNN is proved to show better prediction result than common feedforward Bayesian neural network.

  • PDF

Recurrent Neural Network with Backpropagation Through Time Learning Algorithm for Arabic Phoneme Recognition

  • Ismail, Saliza;Ahmad, Abdul Manan
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2004년도 ICCAS
    • /
    • pp.1033-1036
    • /
    • 2004
  • The study on speech recognition and understanding has been done for many years. In this paper, we propose a new type of recurrent neural network architecture for speech recognition, in which each output unit is connected to itself and is also fully connected to other output units and all hidden units [1]. Besides that, we also proposed the new architecture and the learning algorithm of recurrent neural network such as Backpropagation Through Time (BPTT, which well-suited. The aim of the study was to observe the difference of Arabic's alphabet like "alif" until "ya". The purpose of this research is to upgrade the people's knowledge and understanding on Arabic's alphabet or word by using Recurrent Neural Network (RNN) and Backpropagation Through Time (BPTT) learning algorithm. 4 speakers (a mixture of male and female) are trained in quiet environment. Neural network is well-known as a technique that has the ability to classified nonlinear problem. Today, lots of researches have been done in applying Neural Network towards the solution of speech recognition [2] such as Arabic. The Arabic language offers a number of challenges for speech recognition [3]. Even through positive results have been obtained from the continuous study, research on minimizing the error rate is still gaining lots attention. This research utilizes Recurrent Neural Network, one of Neural Network technique to observe the difference of alphabet "alif" until "ya".

  • PDF

Convolution Neural Network와 Recurrent Neural Network를 활용한 네트워크 패킷 분류 (Network Packet Classification Using Convolution Neural Network and Recurrent Neural Network)

  • 임현교;김주봉;한연희
    • 한국정보처리학회:학술대회논문집
    • /
    • 한국정보처리학회 2018년도 춘계학술발표대회
    • /
    • pp.16-18
    • /
    • 2018
  • 최근 네트워크 상에 새롭고 다양한 어플리케이션들이 생겨나면서 이에 따른 적절한 어플리케이션별 서비스 제공을 위한 패킷 분류 방법이 요구되고 있다. 이로 인하여 딥 러닝 기술이 발전 하면서 이를 이용한 네트워크 트래픽 분류 방법들이 제안되고 있다. 따라서, 본 논문에서는 딥 러닝 기술 중 Convolution Neural Network 와 Recurrent Neural Network 를 동시에 활용한 네트워크 패킷 분류 방법을 제안한다.

오차 자기 순환 신경회로망을 이용한 현가시스템 인식과 슬라이딩 모드 제어기 개발 (Identification of suspension systems using error self recurrent neural network and development of sliding mode controller)

  • 송광현;이창구;김성중
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 1997년도 한국자동제어학술회의논문집; 한국전력공사 서울연수원; 17-18 Oct. 1997
    • /
    • pp.625-628
    • /
    • 1997
  • In this paper the new neural network and sliding mode suspension controller is proposed. That neural network is error self-recurrent neural network. For fast on-line learning, this paper use recursive least squares method. A new neural networks converges considerably faster than the backpropagation algorithm and has advantages of being less affected by the poor initial weights and learning rate. The controller for suspension systems is designed according to sliding mode technique based on new proposed neural network.

  • PDF

칼만필터로 훈련되는 순환신경망을 이용한 시변채널 등화 (Equalization of Time-Varying Channels using a Recurrent Neural Network Trained with Kalman Filters)

  • 최종수;권오신
    • 제어로봇시스템학회논문지
    • /
    • 제9권11호
    • /
    • pp.917-924
    • /
    • 2003
  • Recurrent neural networks have been successfully applied to communications channel equalization. Major disadvantages of gradient-based learning algorithms commonly employed to train recurrent neural networks are slow convergence rates and long training sequences required for satisfactory performance. In a high-speed communications system, fast convergence speed and short training symbols are essential. We propose decision feedback equalizers using a recurrent neural network trained with Kalman filtering algorithms. The main features of the proposed recurrent neural equalizers, utilizing extended Kalman filter (EKF) and unscented Kalman filter (UKF), are fast convergence rates and good performance using relatively short training symbols. Experimental results for two time-varying channels are presented to evaluate the performance of the proposed approaches over a conventional recurrent neural equalizer.