• Title/Summary/Keyword: reinforcement algorithms

Search Result 148, Processing Time 0.02 seconds

A Study on Performance Improvement of Evolutionary Algorithms Using Reinforcement Learning (강화학습을 이용한 진화 알고리즘의 성능개선에 대한 연구)

  • 이상환;심귀보
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 1998.10a
    • /
    • pp.420-426
    • /
    • 1998
  • Evolutionary algorithms are probabilistic optimization algorithms based on the model of natural evolution. Recently the efforts to improve the performance of evolutionary algorithms have been made extensively. In this paper, we introduce the research for improving the convergence rate and search faculty of evolution algorithms by using reinforcement learning. After providing an introduction to evolution algorithms and reinforcement learning, we present adaptive genetic algorithms, reinforcement genetic programming, and reinforcement evolution strategies which are combined with reinforcement learning. Adaptive genetic algorithms generate mutation probabilities of each locus by interacting with the environment according to reinforcement learning. Reinforcement genetic programming executes crossover and mutation operations based on reinforcement and inhibition mechanism of reinforcement learning. Reinforcement evolution strategies use the variances of fitness occurred by mutation to make the reinforcement signals which estimate and control the step length.

  • PDF

Comparison of value-based Reinforcement Learning Algorithms in Cart-Pole Environment

  • Byeong-Chan Han;Ho-Chan Kim;Min-Jae Kang
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.15 no.3
    • /
    • pp.166-175
    • /
    • 2023
  • Reinforcement learning can be applied to a wide variety of problems. However, the fundamental limitation of reinforcement learning is that it is difficult to derive an answer within a given time because the problems in the real world are too complex. Then, with the development of neural network technology, research on deep reinforcement learning that combines deep learning with reinforcement learning is receiving lots of attention. In this paper, two types of neural networks are combined with reinforcement learning and their characteristics were compared and analyzed with existing value-based reinforcement learning algorithms. Two types of neural networks are FNN and CNN, and existing reinforcement learning algorithms are SARSA and Q-learning.

Enhanced Machine Learning Algorithms: Deep Learning, Reinforcement Learning, and Q-Learning

  • Park, Ji Su;Park, Jong Hyuk
    • Journal of Information Processing Systems
    • /
    • v.16 no.5
    • /
    • pp.1001-1007
    • /
    • 2020
  • In recent years, machine learning algorithms are continuously being used and expanded in various fields, such as facial recognition, signal processing, personal authentication, and stock prediction. In particular, various algorithms, such as deep learning, reinforcement learning, and Q-learning, are continuously being improved. Among these algorithms, the expansion of deep learning is rapidly changing. Nevertheless, machine learning algorithms have not yet been applied in several fields, such as personal authentication technology. This technology is an essential tool in the digital information era, walking recognition technology as promising biometrics, and technology for solving state-space problems. Therefore, algorithm technologies of deep learning, reinforcement learning, and Q-learning, which are typical machine learning algorithms in various fields, such as agricultural technology, personal authentication, wireless network, game, biometric recognition, and image recognition, are being improved and expanded in this paper.

Comparison of Reinforcement Learning Algorithms used in Game AI (게임 인공지능에 사용되는 강화학습 알고리즘 비교)

  • Kim, Deokhyung;Jung, Hyunjun
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.693-696
    • /
    • 2021
  • There are various algorithms in reinforcement learning, and the algorithm used differs depending on the field. Even in games, specific algorithms are used when developing AI (artificial intelligence) using reinforcement learning. Different algorithms have different learning methods, so artificial intelligence is created differently. Therefore, the developer has to choose the appropriate algorithm to implement the AI for the purpose. To do that, the developer needs to know the algorithm's learning method and which algorithms are effective for which AI. Therefore, this paper compares the learning methods of three algorithms, SAC, PPO, and POCA, which are algorithms used to implement game AI. These algorithms are practical to apply to which types of AI implementations.

  • PDF

Performance Improvement of Genetic Algorithms by Reinforcement Learning (강화학습을 통한 유전자 알고리즘의 성능개선)

  • 이상환;전효병;심귀보
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 1998.03a
    • /
    • pp.81-84
    • /
    • 1998
  • Genetic Algorithms (GAs) are stochastic algorithms whose search methods model some natural phenomena. The procedure of GAs may be divided into two sub-procedures : Operation and Selection. Chromosomes can produce new offspring by means of operation, and the fitter chromosomes can produce more offspring than the less fit ones by means of selection. However, operation which is executed randomly and has some limits to its execution can not guarantee to produce fitter chromosomes. Thus, we propose a method which gives a directional information to the genetic operator by reinforcement learning. It can be achived by using neural networks to apply reinforcement learning to the genetic operator. We use the amount of fitness change which can be considered as reinforcement signal to calcualte the error terms for the output units. Then the weights are updated using backpropagtion algorithm. The performance improvement of GAs using reinforcement learning can be measured by applying the pr posed method to GA-hard problem.

  • PDF

Comparison of Reinforcement Learning Algorithms for a 2D Racing Game Learning Agent (2D 레이싱 게임 학습 에이전트를 위한 강화 학습 알고리즘 비교 분석)

  • Lee, Dongcheul
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.20 no.1
    • /
    • pp.171-176
    • /
    • 2020
  • Reinforcement learning is a well-known method for training an artificial software agent for a video game. Even though many reinforcement learning algorithms have been proposed, their performance was varies depending on an application area. This paper compares the performance of the algorithms when we train our reinforcement learning agent for a 2D racing game. We defined performance metrics to analyze the results and plotted them into various graphs. As a result, we found ACER (Actor Critic with Experience Replay) achieved the best rewards than other algorithms. There was 157% gap between ACER and the worst algorithm.

Reinforcement Learning Using State Space Compression (상태 공간 압축을 이용한 강화학습)

  • Kim, Byeong-Cheon;Yun, Byeong-Ju
    • The Transactions of the Korea Information Processing Society
    • /
    • v.6 no.3
    • /
    • pp.633-640
    • /
    • 1999
  • Reinforcement learning performs learning through interacting with trial-and-error in dynamic environment. Therefore, in dynamic environment, reinforcement learning method like Q-learning and TD(Temporal Difference)-learning are faster in learning than the conventional stochastic learning method. However, because many of the proposed reinforcement learning algorithms are given the reinforcement value only when the learning agent has reached its goal state, most of the reinforcement algorithms converge to the optimal solution too slowly. In this paper, we present COMREL(COMpressed REinforcement Learning) algorithm for finding the shortest path fast in a maze environment, select the candidate states that can guide the shortest path in compressed maze environment, and learn only the candidate states to find the shortest path. After comparing COMREL algorithm with the already existing Q-learning and Priortized Sweeping algorithm, we could see that the learning time shortened very much.

  • PDF

A Study on Adaptive Random Signal-Based Learning Employing Genetic Algorithms and Simulated Annealing (유전 알고리즘과 시뮬레이티드 어닐링이 적용된 적응 랜덤 신호 기반 학습에 관한 연구)

  • Han, Chang-Wook;Park, Jung-Il
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.7 no.10
    • /
    • pp.819-826
    • /
    • 2001
  • Genetic algorithms are becoming more popular because of their relative simplicity and robustness. Genetic algorithms are global search techniques for nonlinear optimization. However, traditional genetic algorithms, though robust, are generally not the most successful optimization algorithm on any particular domain because they are poor at hill-climbing, whereas simulated annealing has the ability of probabilistic hill-climbing. Therefore, hybridizing a genetic algorithm with other algorithms can produce better performance than using the genetic algorithm or other algorithms independently. In this paper, we propose an efficient hybrid optimization algorithm named the adaptive random signal-based learning. Random signal-based learning is similar to the reinforcement learning of neural networks. This paper describes the application of genetic algorithms and simulated annealing to a random signal-based learning in order to generate the parameters and reinforcement signal of the random signal-based learning, respectively. The validity of the proposed algorithm is confirmed by applying it to two different examples.

  • PDF

New reinforcement algorithms in discontinuous deformation analysis for rock failure

  • Chen, Yunjuan;Zhu, Weishen;Li, Shucai;Zhang, Xin
    • Geomechanics and Engineering
    • /
    • v.11 no.6
    • /
    • pp.787-803
    • /
    • 2016
  • DDARF (Discontinuous Deformation Analysis for Rock Failure) is a numerical algorithm for simulating jointed rock masses' discontinuous deformation. While its reinforcement simulation is only limited to end-anchorage bolt, which is assumed to be a linear spring simply. Here, several new reinforcement modes in DDARF are proposed, including lining reinforcement, full-length anchorage bolt and equivalent reinforcement. In the numerical simulation, lining part is assigned higher mechanical strength than surrounding rock masses, it may include multiple virtual joints or not, depending on projects. There must be no embedding or stretching between lining blocks and surrounding blocks. To realize simulation of the full-length anchorage bolt, at every discontinuity passed through the bolt, a set of normal and tangential spring needs to be added along the bolt's axial and tangential direction. Thus, bolt's axial force, shearing force and full-length anchorage effect are all realized synchronously. And, failure criterions of anchorage effect are established for different failure modes. In the meantime, from the perspective of improving surrounding rock masses' overall strength, a new equivalent and tentative simulation method is proposed, it can save calculation storage and improve efficiency. Along the text, simulation algorithms and applications of these new reinforcement modes in DDARF are given.

PESA: Prioritized experience replay for parallel hybrid evolutionary and swarm algorithms - Application to nuclear fuel

  • Radaideh, Majdi I.;Shirvan, Koroush
    • Nuclear Engineering and Technology
    • /
    • v.54 no.10
    • /
    • pp.3864-3877
    • /
    • 2022
  • We propose a new approach called PESA (Prioritized replay Evolutionary and Swarm Algorithms) combining prioritized replay of reinforcement learning with hybrid evolutionary algorithms. PESA hybridizes different evolutionary and swarm algorithms such as particle swarm optimization, evolution strategies, simulated annealing, and differential evolution, with a modular approach to account for other algorithms. PESA hybridizes three algorithms by storing their solutions in a shared replay memory, then applying prioritized replay to redistribute data between the integral algorithms in frequent form based on their fitness and priority values, which significantly enhances sample diversity and algorithm exploration. Additionally, greedy replay is used implicitly to improve PESA exploitation close to the end of evolution. PESA features in balancing exploration and exploitation during search and the parallel computing result in an agnostic excellent performance over a wide range of experiments and problems presented in this work. PESA also shows very good scalability with number of processors in solving an expensive problem of optimizing nuclear fuel in nuclear power plants. PESA's competitive performance and modularity over all experiments allow it to join the family of evolutionary algorithms as a new hybrid algorithm; unleashing the power of parallel computing for expensive optimization.