• 제목/요약/키워드: Reinforcement neural network

검색결과 130건 처리시간 0.03초

Traffic-based reinforcement learning with neural network algorithm in fog computing environment

  • Jung, Tae-Won;Lee, Jong-Yong;Jung, Kye-Dong
    • International Journal of Internet, Broadcasting and Communication
    • /
    • 제12권1호
    • /
    • pp.144-150
    • /
    • 2020
  • Reinforcement learning is a technology that can present successful and creative solutions in many areas. This reinforcement learning technology was used to deploy containers from cloud servers to fog servers to help them learn the maximization of rewards due to reduced traffic. Leveraging reinforcement learning is aimed at predicting traffic in the network and optimizing traffic-based fog computing network environment for cloud, fog and clients. The reinforcement learning system collects network traffic data from the fog server and IoT. Reinforcement learning neural networks, which use collected traffic data as input values, can consist of Long Short-Term Memory (LSTM) neural networks in network environments that support fog computing, to learn time series data and to predict optimized traffic. Description of the input and output values of the traffic-based reinforcement learning LSTM neural network, the composition of the node, the activation function and error function of the hidden layer, the overfitting method, and the optimization algorithm.

강화 학습에 기반한 뉴럴-퍼지 제어기 (Neural-Fuzzy Controller Based on Reinforcement Learning)

  • 박영철;김대수;심귀보
    • 한국지능시스템학회:학술대회논문집
    • /
    • 한국퍼지및지능시스템학회 2000년도 춘계학술대회 학술발표 논문집
    • /
    • pp.245-248
    • /
    • 2000
  • 본 논문에서는 강화 학습 개념을 도입하여 자율이동 로봇의 성능을 개선하고자 한다. 본 논문에서 사용되는 시스템은 크게 두 부분으로 나눌 수가 있다. 즉, 뉴럴 퍼지 부분과 동적귀환 신경회로망이다. 뉴럴 퍼지 부분은 로봇의 다음 행동을 결정하는 부분이다. 또한 동적귀환 신경회로망으로부터 내부 강화 신호를 받아 학습을 하여 최적의 행동을 결정하게 된다. 동적 귀환신경회로망은 환경으로부터 외부 강화신호를 입력으로 받아 뉴럴 퍼지의 행동결정에 대해 평가를 한다. 또한 내부강화 신호 값을 결정하는 동적 귀환 신경회로망의 웨이트는 유전자 알고리즘에 의해 진화를 한다. 제안한 알고리즘 구조를 컴퓨터 시뮬레이션상에서 자율 이동 로봇의 제어에 적용을 함으로서 그 유효성을 증명하고자 한다.

  • PDF

Reinforcement Learning Control using Self-Organizing Map and Multi-layer Feed-Forward Neural Network

  • Lee, Jae-Kang;Kim, Il-Hwan
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2003년도 ICCAS
    • /
    • pp.142-145
    • /
    • 2003
  • Many control applications using Neural Network need a priori information about the objective system. But it is impossible to get exact information about the objective system in real world. To solve this problem, several control methods were proposed. Reinforcement learning control using neural network is one of them. Basically reinforcement learning control doesn't need a priori information of objective system. This method uses reinforcement signal from interaction of objective system and environment and observable states of objective system as input data. But many methods take too much time to apply to real-world. So we focus on faster learning to apply reinforcement learning control to real-world. Two data types are used for reinforcement learning. One is reinforcement signal data. It has only two fixed scalar values that are assigned for each success and fail state. The other is observable state data. There are infinitive states in real-world system. So the number of observable state data is also infinitive. This requires too much learning time for applying to real-world. So we try to reduce the number of observable states by classification of states with Self-Organizing Map. We also use neural dynamic programming for controller design. An inverted pendulum on the cart system is simulated. Failure signal is used for reinforcement signal. The failure signal occurs when the pendulum angle or cart position deviate from the defined control range. The control objective is to maintain the balanced pole and centered cart. And four states that is, position and velocity of cart, angle and angular velocity of pole are used for state signal. Learning controller is composed of serial connection of Self-Organizing Map and two Multi-layer Feed-Forward Neural Networks.

  • PDF

ON THE STRUCTURE AND LEARNING OF NEURAL-NETWORK-BASED FUZZY LOGIC CONTROL SYSTEMS

  • C.T. Lin;Lee, C.S. George
    • 한국지능시스템학회:학술대회논문집
    • /
    • 한국퍼지및지능시스템학회 1993년도 Fifth International Fuzzy Systems Association World Congress 93
    • /
    • pp.993-996
    • /
    • 1993
  • This paper addresses the structure and its associated learning algorithms of a feedforward multi-layered connectionist network, which has distributed learning abilities, for realizing the basic elements and functions of a traditional fuzzy logic controller. The proposed neural-network-based fuzzy logic control system (NN-FLCS) can be contrasted with the traditional fuzzy logic control system in their network structure and learning ability. An on-line supervised structure/parameter learning algorithm dynamic learning algorithm can find proper fuzzy logic rules, membership functions, and the size of output fuzzy partitions simultaneously. Next, a Reinforcement Neural-Network-Based Fuzzy Logic Control System (RNN-FLCS) is proposed which consists of two closely integrated Neural-Network-Based Fuzzy Logic Controllers (NN-FLCS) for solving various reinforcement learning problems in fuzzy logic systems. One NN-FLC functions as a fuzzy predictor and the other as a fuzzy controller. As ociated with the proposed RNN-FLCS is the reinforcement structure/parameter learning algorithm which dynamically determines the proper network size, connections, and parameters of the RNN-FLCS through an external reinforcement signal. Furthermore, learning can proceed even in the period without any external reinforcement feedback.

  • PDF

신경회로망을 이용한 도립전자의 학습제어 (Learning Control of Inverted Pendulum Using Neural Networks)

  • 이재강;김일환
    • 산업기술연구
    • /
    • 제24권A호
    • /
    • pp.99-107
    • /
    • 2004
  • This paper considers reinforcement learning control with the self-organizing map. Reinforcement learning uses the observable states of objective system and signals from interaction of the system and the environments as input data. For fast learning in neural network training, it is necessary to reduce learning data. In this paper, we use the self-organizing map to parition the observable states. Partitioning states reduces the number of learning data which is used for training neural networks. And neural dynamic programming design method is used for the controller. For evaluating the designed reinforcement learning controller, an inverted pendulum of the cart system is simulated. The designed controller is composed of serial connection of self-organizing map and two Multi-layer Feed-Forward Neural Networks.

  • PDF

Hand Reaching Movement Acquired through Reinforcement Learning

  • Shibata, Katsunari;Sugisaka, Masanori;Ito, Koji
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2000년도 제15차 학술회의논문집
    • /
    • pp.474-474
    • /
    • 2000
  • This paper shows that a system with two-link arm can obtain hand reaching movement to a target object projected on a visual sensor by reinforcement learning using a layered neural network. The reinforcement signal, which is an only signal from the environment, is given to the system only when the hand reaches the target object. The neural network computes two joint torques from visual sensory signals, joint angles, and joint angular velocities considering the urn dynamics. It is known that the trajectory of the voluntary movement o( human hand reaching is almost straight, and the hand velocity changes like bell-shape. Although there are some exceptions, the properties of the trajectories obtained by the reinforcement learning are somewhat similar to the experimental result of the human hand reaching movement.

  • PDF

이중 학습에 의한 선형동기모터의 위치제어 (Position Control of Linear Synchronous Motor by Dual Learning)

  • 박정일;서성호;울루구벡
    • 한국정밀공학회지
    • /
    • 제29권1호
    • /
    • pp.79-86
    • /
    • 2012
  • This paper proposes PID and RIC (Robust Internal-loop Compensator) based motion controller using dual learning algorithm for position control of linear synchronous motor respectively. Its gains are auto-tuned by using two learning algorithms, reinforcement learning and neural network. The feedback controller gains are tuned by reinforcement learning, and then the feedforward controller gains are tuned by neural network. Experiments prove the validity of dual learning algorithm. The RIC controller has better performance than does the PID-feedforward controller in reducing tracking error and disturbance rejection. Neural network shows its ability to decrease tracking error and to reject disturbance in the stop range of the target position and home.

인공신경망과 유전 알고리즘을 이용한 팩맨 게임 강화학습 (Pacman Game Reinforcement Learning Using Artificial Neural-network and Genetic Algorithm)

  • 박진수;이호정;황두연;조수선
    • 대한임베디드공학회논문지
    • /
    • 제15권5호
    • /
    • pp.261-268
    • /
    • 2020
  • Genetic algorithms find the optimal solution by mimicking the evolution of natural organisms. In this study, the genetic algorithm was used to enable Pac-Man's reinforcement learning, and a simulator to observe the evolutionary process was implemented. The purpose of this paper is to reinforce the learning of the Pacman AI of the simulator, and utilize genetic algorithm and artificial neural network as the method. In particular, by building a low-power artificial neural network and applying it to a genetic algorithm, it was intended to increase the possibility of implementation in a low-power embedded system.

Learning Less Random to Learn Better in Deep Reinforcement Learning with Noisy Parameters

  • Kim, Chayoung
    • 한국정보기술학회 영문논문지
    • /
    • 제9권1호
    • /
    • pp.127-134
    • /
    • 2019
  • In terms of deep Reinforcement Learning (RL), exploration can be worked stochastically in the action of a state space. On the other hands, exploitation can be done the proportion of well generalization behaviors. The balance of exploration and exploitation is extremely important for better results. The randomly selected action with ε-greedy for exploration has been regarded as a de facto method. There is an alternative method to add noise parameters into a neural network for richer exploration. However, it is not easy to predict or detect over-fitting with the stochastically exploration in the perturbed neural network. Moreover, the well-trained agents in RL do not necessarily prevent or detect over-fitting in the neural network. Therefore, we suggest a novel design of a deep RL by the balance of the exploration with drop-out to reduce over-fitting in the perturbed neural networks.

Comparison of value-based Reinforcement Learning Algorithms in Cart-Pole Environment

  • Byeong-Chan Han;Ho-Chan Kim;Min-Jae Kang
    • International Journal of Internet, Broadcasting and Communication
    • /
    • 제15권3호
    • /
    • pp.166-175
    • /
    • 2023
  • Reinforcement learning can be applied to a wide variety of problems. However, the fundamental limitation of reinforcement learning is that it is difficult to derive an answer within a given time because the problems in the real world are too complex. Then, with the development of neural network technology, research on deep reinforcement learning that combines deep learning with reinforcement learning is receiving lots of attention. In this paper, two types of neural networks are combined with reinforcement learning and their characteristics were compared and analyzed with existing value-based reinforcement learning algorithms. Two types of neural networks are FNN and CNN, and existing reinforcement learning algorithms are SARSA and Q-learning.