• 제목/요약/키워드: Learning control gain

검색결과 87건 처리시간 0.023초

Estimation of learning gain in iterative learning control using neural networks

  • Choi, Jin-Young;Park, Hyun-Joo
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 1996년도 Proceedings of the Korea Automatic Control Conference, 11th (KACC); Pohang, Korea; 24-26 Oct. 1996
    • /
    • pp.91-94
    • /
    • 1996
  • This paper presents an approach to estimation of learning gain in iterative learning control for discrete-time affine nonlinear systems. In iterative learning control, to determine learning gain satisfying the convergence condition, we have to know the system model. In the proposed method, the input-output equation of a system is identified by neural network refered to as Piecewise Linearly Trained Network (PLTN). Then from the input-output equation, the learning gain in iterative learning law is estimated. The validity of our method is demonstrated by simulations.

  • PDF

Q-Learning을 사용한 로봇팔의 SMCSPO 게인 튜닝 (Gain Tuning for SMCSPO of Robot Arm with Q-Learning)

  • 이진혁;김재형;이민철
    • 로봇학회논문지
    • /
    • 제17권2호
    • /
    • pp.221-229
    • /
    • 2022
  • Sliding mode control (SMC) is a robust control method to control a robot arm with nonlinear properties. A high switching gain of SMC causes chattering problems, although the SMC allows the adequate control performance by giving high switching gain, without the exact robot model containing nonlinear and uncertainty terms. In order to solve this problem, SMC with sliding perturbation observer (SMCSPO) has been researched, where the method can reduce the chattering by compensating the perturbation, which is estimated by the observer, and then choosing a lower switching control gain of SMC. However, optimal gain tuning is necessary to get a better tracking performance and reducing a chattering. This paper proposes a method that the Q-learning automatically tunes the control gains of SMCSPO with an iterative operation. In this tuning method, the rewards of reinforcement learning (RL) are set minus tracking errors of states, and the action of RL is a change of control gain to maximize rewards whenever the iteration number of movements increases. The simple motion test for a 7-DOF robot arm was simulated in MATLAB program to prove this RL tuning algorithm. The simulation showed that this method can automatically tune the control gains for SMCSPO.

일정적응 이득과 이진 강화함수를 갖는 경쟁 학습 신경회로망 (Competitive Learning Neural Network with Binary Reinforcement and Constant Adaptation Gain)

  • 석진욱;조성원;최경삼
    • 대한전기학회:학술대회논문집
    • /
    • 대한전기학회 1994년도 추계학술대회 논문집 학회본부
    • /
    • pp.326-328
    • /
    • 1994
  • A modified Kohonen's simple Competitive Learning(SCL) algorithm which has binary reinforcement function and a constant adaptation gain is proposed. In contrast to the time-varing adaptation gain of the original Kohonen's SCL algorithm, the proposed algorithm uses a constant adaptation gain, and adds a binary reinforcement function in order to compensate for the lowered learning ability of SCL due to the constant adaptation gain. Since the proposed algorithm does not have the complicated multiplication, it's digital hardware implementation is much easier than one of the original SCL.

  • PDF

학습이득 조절기에 의한 직류 모터 속도제어 (D.C. Motor Speed Control by Learning Gain Regulator)

  • 박왈서;이성수;김용욱
    • 조명전기설비학회논문지
    • /
    • 제19권6호
    • /
    • pp.82-86
    • /
    • 2005
  • PID 제어기는 산업자동화 설비에 널리 쓰이고 있다. 하지만 시스템 특성이 간헐 또는 연속적으로 변화할 때에 정밀제어를 위한 새로운 매개변수 결정이 쉽지 않다. 이를 해결하기 위한 방법으로 본 논문에서는 PID 제어기와 같은 기능을 갖는 학습이득조절기를 제안하였다. 시스템의 적절한 학습이득은 델타 학습규칙에 의해서 결정된다. 제안된 학습이득 조절기의 기능은 직류 전동기의 모의실험에 의해 확인하였다.

기계학습 알고리즘을 이용한 UAS 제어계수 실시간 자동 조정 시스템 (UAS Automatic Control Parameter Tuning System using Machine Learning Module)

  • 문미선;송강;송동호
    • 한국항행학회논문지
    • /
    • 제14권6호
    • /
    • pp.874-881
    • /
    • 2010
  • 무인기의 자동 비행 제어 시스템은 기체의 형태, 크기, 무게 등의 정적 및 동적 변화에 따라 스스로 비행계수를 조정하여 목표 비행궤적을 정확히 따라가도록 제어할 필요가 있다. 본 논문에서는 PID 제어 기법을 이용하는 비행제어시스템에 기계학습모듈(MLM)을 추가하여 기체의 특성 변화에 따라 제어계수를 비행중 실시간 자동으로 조정하는 시스템을 제안한다. MLM은 선형회귀분석과 보정학습을 이용하여 설계되었으며 MLM을 통해 학습된 제어계수의 적합성을 평가하는 평가모듈(EvM)을 함께 모델링 하였다. 이 시스템은 FDC 비버 시뮬레이터를 기반으로 실험하였으며 그 결과를 분석 제시하였다.

A general dynamic iterative learning control scheme with high-gain feedback

  • Kuc, Tae-Yong;Nam, Kwanghee
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 1989년도 한국자동제어학술회의논문집; Seoul, Korea; 27-28 Oct. 1989
    • /
    • pp.1140-1145
    • /
    • 1989
  • A general dynamic iterative learning control scheme is proposed for a class of nonlinear systems. Relying on stabilizing high-gain feedback loop, it is possible to show the existence of Cauchy sequence of feedforward control input error with iteration numbers, which results in a uniform convergance of system state trajectory to the desired one.

  • PDF

Fuzzy Gain Scheduling of Velocity PI Controller with Intelligent Learning Algorithm for Reactor Control

  • Kim, Dong-Yun;Seong, Poong-Hyun
    • 한국원자력학회:학술대회논문집
    • /
    • 한국원자력학회 1996년도 추계학술발표회논문집(1)
    • /
    • pp.73-78
    • /
    • 1996
  • In this study, we proposed a fuzzy gain scheduler with intelligent learning algorithm for a reactor control. In the proposed algorithm, we used the gradient descent method to learn the rule bases of a fuzzy algorithm. These rule bases are learned toward minimizing an objective function, which is called a performance cost function. The objective of fuzzy gain scheduler with intelligent learning algorithm is the generation of adequate gains, which minimize the error of system. The condition of every plant is generally changed as time gose. That is, the initial gains obtained through the analysis of system are no longer suitable for the changed plant. And we need to set new gains, which minimize the error stemmed from changing the condition of a plant. In this paper, we applied this strategy for reactor control of nuclear power plant (NPP), and the results were compared with those of a simple PI controller, which has fixed gains. As a result, it was shown that the proposed algorithm was superior to the simple PI controller.

  • PDF

신경망을 이용한 PID 제어기의 제어 사양 최적의 이득값 추정 (Optimal Condition Gain Estimation of PID Controller using Neural Networks)

  • 손준혁;서보혁
    • 대한전기학회:학술대회논문집
    • /
    • 대한전기학회 2003년도 학술회의 논문집 정보 및 제어부문 B
    • /
    • pp.717-719
    • /
    • 2003
  • Recently Neural Network techniques have widely used in adaptive and learning control schemes for production systems. However, generally it costs a lot of time for learning in the case applied in control system. Furthermore, the physical meaning of neural networks constructed as a result is not obvious. And in practice since it is difficult to the PID gains suitably lots of researches have been reported with respect to turning schemes of PID gains. A Neural Network-based PID control scheme is proposed, which extracts skills of human experts as PID gains. This controller is designed by using three-layered neural networks. The effectiveness of the proposed Neural Network-based PID control scheme is investigated through an application for a production control system. This control method can enable a plant to operate smoothy and obviously as the plant condition varies with any unexpected accident.

  • PDF

A Neurofuzzy Algorithm-Based Advanced Bilateral Controller for Telerobot Systems

  • Cha, Dong-hyuk;Cho, Hyung-Suck
    • Transactions on Control, Automation and Systems Engineering
    • /
    • 제4권1호
    • /
    • pp.100-107
    • /
    • 2002
  • The advanced bilateral control algorithm, which can enlarge a reflected force by combining force reflection and compliance control, greatly enhances workability in teleoperation. In this scheme the maximum boundaries of a compliance controller and a force reflection gain guaranteeing stability and good task performance greatly depend upon characteristics of a slave arm, a master arm, and an environment. These characteristics, however, are generally unknown in teleoperation. It is, therefore, very difficult to determine such maximum boundary of the gain. The paper presented a novel method for design of an advanced bilateral controller. The factors affecting task performance and stability in the advanced bilateral controller were analyzed and a design guideline was presented. The neurofuzzy compliance model (NFCM)-based bilateral control proposed herein is an algorithm designed to automatically determine the suitable compliance for a given task or environment. The NFCM, composed of a fuzzy logic controller (FLC) and a rule-learning mechanism, is used as a compliance controller. The FLC generates compliant motions according to contact forces. The rule-learning mechanism, which is based upon the reinforcement learning algorithm, trains the rule-base of the FLC until the given task is done successfully. Since the scheme allows the use of large force reflection gain, it can assure good task performance. Moreover, the scheme does not require any priori knowledge on a slave arm dynamics, a slave arm controller and an environment, and thus, it can be easily applied to the control of any telerobot systems. Through a series of experiments effectiveness of the proposed algorithm has been verified.

PID 학습제어기를 이용한 가변부하 직류서보전동기의 실시간 제어 (Real-Time Control of DC Sevo Motor with Variable Load Using PID-Learning Controller)

  • 김상훈;정인석;강영호;남문현;김낙교
    • 대한전기학회논문지:시스템및제어부문D
    • /
    • 제50권3호
    • /
    • pp.107-113
    • /
    • 2001
  • This paper deals with speed control of DC servo motor using a PID controller with a gain tuning based on a Back-Propagation(BP) Learning Algorithm. Conventionally a PID controller has been used in the industrial control. But a PID controller should produce suitable parameters for each system. Also, variables of the PID controller should be changed according to environments, disturbances and loads. In this paper described by a experiment that contained a method using a PID controller with a gain tuning based on a Back-Propagation(BP) Learning Algorithm, we developed speed characteristics of a DC servo motor on variable loads. The parameters of the controller are determined by neural network performed on on-line system after training the neural network on off-line system.

  • PDF