• Title/Summary/Keyword: Learning Control Algorithm

Search Result 947, Processing Time 0.029 seconds

An Improved Domain-Knowledge-based Reinforcement Learning Algorithm

  • Jang, Si-Young;Suh, Il-Hong
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2003.10a
    • /
    • pp.1309-1314
    • /
    • 2003
  • If an agent has a learning ability using previous knowledge, then it is expected that the agent can speed up learning by interacting with environment. In this paper, we present an improved reinforcement learning algorithm using domain knowledge which can be represented by problem-independent features and their classifiers. Here, neural networks are employed as knowledge classifiers. To show the validity of our proposed algorithm, computer simulations are illustrated, where navigation problem of a mobile robot and a micro aerial vehicle(MAV) are considered.

  • PDF

The Self-tuning PID Control Based on Real-time Adaptive Learning Evolutionary Algorithm (실시간 적응 학습 진화 알고리듬을 이용한 자기 동조 PID 제어)

  • Chang, Sung-Ouk;Lee, Jin-Kul
    • Transactions of the Korean Society of Mechanical Engineers A
    • /
    • v.27 no.9
    • /
    • pp.1463-1468
    • /
    • 2003
  • This paper presented the real-time self-tuning learning control based on evolutionary computation, which proves its superiority in finding of the optimal solution at the off-line learning method. The individuals of the populations are reduced in order to learn the evolutionary strategy in real-time, and new method that guarantee the convergence of evolutionary mutations is proposed. It is possible to control the control object slightly varied as time changes. As the state value of the control object is generated, evolutionary strategy is applied each sampling time because the learning process of an estimation, selection, mutation is done in real-time. These algorithms can be applied; the people who do not have knowledge about the technical tuning of dynamic systems could design the controller or problems in which the characteristics of the system dynamics are slightly varied as time changes.

Avoidance Behavior of Small Mobile Robots based on the Successive Q-Learning

  • Kim, Min-Soo
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2001.10a
    • /
    • pp.164.1-164
    • /
    • 2001
  • Q-learning is a recent reinforcement learning algorithm that does not need a modeling of environment and it is a suitable approach to learn behaviors for autonomous agents. But when it is applied to multi-agent learning with many I/O states, it is usually too complex and slow. To overcome this problem in the multi-agent learning system, we propose the successive Q-learning algorithm. Successive Q-learning algorithm divides state-action pairs, which agents can have, into several Q-functions, so it can reduce complexity and calculation amounts. This algorithm is suitable for multi-agent learning in a dynamically changing environment. The proposed successive Q-learning algorithm is applied to the prey-predator problem with the one-prey and two-predators, and its effectiveness is verified from the efficient avoidance ability of the prey agent.

  • PDF

A Neurofuzzy Algorithm-Based Advanced Bilateral Controller for Telerobot Systems

  • Cha, Dong-hyuk;Cho, Hyung-Suck
    • Transactions on Control, Automation and Systems Engineering
    • /
    • v.4 no.1
    • /
    • pp.100-107
    • /
    • 2002
  • The advanced bilateral control algorithm, which can enlarge a reflected force by combining force reflection and compliance control, greatly enhances workability in teleoperation. In this scheme the maximum boundaries of a compliance controller and a force reflection gain guaranteeing stability and good task performance greatly depend upon characteristics of a slave arm, a master arm, and an environment. These characteristics, however, are generally unknown in teleoperation. It is, therefore, very difficult to determine such maximum boundary of the gain. The paper presented a novel method for design of an advanced bilateral controller. The factors affecting task performance and stability in the advanced bilateral controller were analyzed and a design guideline was presented. The neurofuzzy compliance model (NFCM)-based bilateral control proposed herein is an algorithm designed to automatically determine the suitable compliance for a given task or environment. The NFCM, composed of a fuzzy logic controller (FLC) and a rule-learning mechanism, is used as a compliance controller. The FLC generates compliant motions according to contact forces. The rule-learning mechanism, which is based upon the reinforcement learning algorithm, trains the rule-base of the FLC until the given task is done successfully. Since the scheme allows the use of large force reflection gain, it can assure good task performance. Moreover, the scheme does not require any priori knowledge on a slave arm dynamics, a slave arm controller and an environment, and thus, it can be easily applied to the control of any telerobot systems. Through a series of experiments effectiveness of the proposed algorithm has been verified.

Design of multivariable learning controller in frequency domain (주파수 영역에서 다변수 학습제어기의 설계)

  • 김원철;조진원;이광순
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1993.10a
    • /
    • pp.760-765
    • /
    • 1993
  • A multivariable learning control is designed in frequency domain. A general to of feedback assisted learning scheme is considered and an inverse model based learning algorithm is derived through convergence analysis in frequency domain. Performance of the proposed control method is evaluated through numerical simulation.

  • PDF

Q-learning for intersection traffic flow Control based on agents

  • Zhou, Xuan;Chong, Kil-To
    • Proceedings of the IEEK Conference
    • /
    • 2009.05a
    • /
    • pp.94-96
    • /
    • 2009
  • In this paper, we present the Q-learning method for adaptive traffic signal control on the basis of multi-agent technology. The structure is composed of sixphase agents and one intersection agent. Wireless communication network provides the possibility of the cooperation of agents. As one kind of reinforcement learning, Q-learning is adopted as the algorithm of the control mechanism, which can acquire optical control strategies from delayed reward; furthermore, we adopt dynamic learning method instead of static method, which is more practical. Simulation result indicates that it is more effective than traditional signal system.

  • PDF

Self-Organizing Fuzzy Modeling Based on Hyperplane-Shaped Clusters (다차원 평면 클러스터를 이용한 자기 구성 퍼지 모델링)

  • Koh, Taek-Beom
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.7 no.12
    • /
    • pp.985-992
    • /
    • 2001
  • This paper proposes a self-organizing fuzzy modeling(SOFUM)which an create a new hyperplane shaped cluster and adjust parameters of the fuzzy model in repetition. The suggested algorithm SOFUM is composed of four steps: coarse tuning. fine tuning cluster creation and optimization of learning rates. In the coarse tuning fuzzy C-regression model(FCRM) clustering and weighted recursive least squared (WRLS) algorithm are used and in the fine tuning gradient descent algorithm is used to adjust parameters of the fuzzy model precisely. In the cluster creation, a new hyperplane shaped cluster is created by applying multiple regression to input/output data with relatively large fuzzy entropy based on parameter tunings of fuzzy model. And learning rates are optimized by utilizing meiosis-genetic algorithm in the optimization of learning rates To check the effectiveness of the suggested algorithm two examples are examined and the performance of the identified fuzzy model is demonstrated via computer simulation.

  • PDF

Neuro-Fuzzy Algorithm for Nuclear Reactor Power Control : Part I

  • Chio, Jung-In;Hah, Yung-Joon
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.5 no.3
    • /
    • pp.52-63
    • /
    • 1995
  • A neuro-fuzzy algorithm is presented for nuclear reactor power control in a pressurized water reactor. Automatic reacotr power control is complicated by the use of control rods because of highly nonlinear dynamics in the axial power shape. Thus, manual shaped controls are usually employed even for the limited capability during the power maneuvers. In an attempt to achieve automatic shape control, a neuro-fuzzy approach is considered because fuzzy algorithms are good at various aspects of operator's knowledge representation while neural networks are efficinet structures capable of learning from experience and adaptation to a changing nuclear core state. In the proposed neuro-fuzzy control scheme, the rule base is formulated based ona multi-input multi-output system and the dynamic back-propagation is used for learning. The neuro-fuzzy powere control algorithm has been tested using simulation fesponses of a Korean standard pressurized water reactor. The results illustrate that the proposed control algorithm would be a parctical strategy for automatic nuclear reactor power control.

  • PDF

Control for crane's swing using fuzzy learning method (퍼지 학습법을 이용한 crane의 과도 진동 제어)

  • 임윤규;정병묵
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1997.10a
    • /
    • pp.450-453
    • /
    • 1997
  • An active control for the swing of crane systems is very important for increasing the productivity. This article introduces the control for the position and the swing of a crane using the fuzzy learning method. Because the crane is a multi-variable system, learning is done to control both position and swing of the crane. Also the fuzzy control rules are separately acquired with the loading and unloading situation of the crane for more accurate control. The result of simulations shows that the crane is just controlled for a very large swing angle of 1 radian within nearly one cycle.

  • PDF

Optimal Design of Magnetic Levitation Controller Using Advanced Teaching-Learning Based Optimization (개선된 수업-학습기반 최적화 알고리즘을 이용한 자기부상 제어기의 최적 설계)

  • Cho, Jae-Hoon;Kim, Yong-Tae
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.64 no.1
    • /
    • pp.90-98
    • /
    • 2015
  • In this paper, an advanced teaching-learning based optimization(TLBO) method for the magnetic levitation controller of Maglev transportation system is proposed to optimize the control performances. An attraction-type levitation system is intrinsically unstable and requires a delicate control. It is difficult to completely satisfy the desired performance through the methods using conventional methods and intelligent optimizations. In the paper, we use TLBO and clonal selection algorithm to choose the optimal control parameters for the magnetic levitation controller. To verify the proposed algorithm, we compare control performances of the proposed method with the genetic algorithm and the particle swarm optimization. The simulation results show that the proposed method is more effective than conventional methods.