• Title/Summary/Keyword: Learning Control Algorithm

Search Result 944, Processing Time 0.031 seconds

Robot learning control with fast convergence (빠른 수렴성을 갖는 로보트 학습제어)

  • 양원영;홍호선
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1988.10a
    • /
    • pp.67-71
    • /
    • 1988
  • We present an algorithm that uses trajectory following errors to improve a feedforward command to a robot in the iterative manner. It has been shown that when the manipulator handles an unknown object, the P-type learning algorithm can make the trajectory converge to a desired path and also that the proposed learning control algorithm performs better than the other type learning control algorithm. A numerical simulation of a three degree of freedom manipulator such as PUMA-560 ROBOT has been performed to illustrate the effectiveness of the proposed learning algorithm.

  • PDF

Evolutionary Learning-Rate Selection for BPNN with Window Control Scheme

  • Hoon, Jung-Sung
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 1997.10a
    • /
    • pp.301-308
    • /
    • 1997
  • The learning speed of the neural networks, the most important factor in applying to real problems, greatly depends on the learning rate of the networks, Three approaches-empirical, deterministic, and stochastic ones-have been proposed to date. We proposed a new learning-rate selection algorithm using an evolutionary programming search scheme. Even though the performance of our method showed better than those of the other methods, it was found that taking much time for selecting evolutionary learning rates made the performance of our method degrade. This was caused by using static intervals (called static windows) in order to update learning rates. Out algorithm with static windows updated the learning rates showed good performance or didn't update the learning rates even though previously updated learning rates shoved bad performance. This paper introduce a window control scheme to avoid such problems. With the window control scheme, our algorithm try to update the learning ra es only when the learning performance is continuously bad during a specified interval. If previously selected learning rates show good performance, new algorithm will not update the learning rates. This diminish the updating time of learning rates greatly. As a result, our algorithm with the window control scheme show better performance than that with static windows. In this paper, we will describe the previous and new algorithm and experimental results.

  • PDF

A Study on Implementation of a Real Time Learning Controller for Direct Drive Manipulator (직접 구동형 매니퓰레이터를 위한 학습 제어기의 실시간 구현에 관한 연구)

  • Jeon, Jong-Wook;An, Hyun-Sik;Lim, Mee-Seub;Kim, Kwon-Ho;Kim, Kwang-Bae;Lee, Kwae-Hi
    • Proceedings of the KIEE Conference
    • /
    • 1993.07a
    • /
    • pp.369-372
    • /
    • 1993
  • In this thesis, we consider an iterative learning controller to control the continuous trajectory of 2 links direct drive robot manipulator and process computer simulation and real-time experiment. To improve control performance, we adapt an iterative learning control algorithm, drive a sufficient condition for convergence from which is drived extended conventional control algorithm and get better performance by extended learning control algorithm than that by conventional algorithm from simulation results. Also, experimental results show that better performance is taken by extended learning algorithm.

  • PDF

Model-based iterative learning control with quadratic criterion for linear batch processes (선형 회분식 공정을 위한 이차 성능 지수에 의한 모델 기반 반복 학습 제어)

  • Lee, Kwang-Soon;Kim, Won-Cheol;Lee, Jay-H
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.2 no.3
    • /
    • pp.148-157
    • /
    • 1996
  • Availability of input trajectories corresponding to desired output trajectories is often important in designing control systems for batch and other transient processes. In this paper, we propose a predictive control-type model-based iterative learning algorithm which is applicable to finding the nominal input trajectories of a linear time-invariant batch process. Unlike the other existing learning control algorithms, the proposed algorithm can be applied to nonsquare systems and has an ability to adjust noise sensitivity as well as convergence rate. A simple model identification technique with which performance of the proposed learning algorithm can be significantly enhanced is also proposed. Performance of the proposed learning algorithm is demonstrated through numerical simulations.

  • PDF

Multi-Channel Speech Enhancement Algorithm Using DOA-based Learning Rate Control (DOA 기반 학습률 조절을 이용한 다채널 음성개선 알고리즘)

  • Kim, Su-Hwan;Lee, Young-Jae;Kim, Young-Il;Jeong, Sang-Bae
    • Phonetics and Speech Sciences
    • /
    • v.3 no.3
    • /
    • pp.91-98
    • /
    • 2011
  • In this paper, a multi-channel speech enhancement method using the linearly constrained minimum variance (LCMV) algorithm and a variable learning rate control is proposed. To control the learning rate for adaptive filters of the LCMV algorithm, the direction of arrival (DOA) is measured for each short-time input signal and the likelihood function of the target speech presence is estimated to control the filter learning rate. Using the likelihood measure, the learning rate is increased during the pure noise interval and decreased during the target speech interval. To optimize the parameter of the mapping function between the likelihood value and the corresponding learning rate, an exhaustive search is performed using the Bark's scale distortion (BSD) as the performance index. Experimental results show that the proposed algorithm outperforms the conventional LCMV with fixed learning rate in the BSD by around 1.5 dB.

  • PDF

Optimal Learning of Fuzzy Neural Network Using Particle Swarm Optimization Algorithm

  • Kim, Dong-Hwa;Cho, Jae-Hoon
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.421-426
    • /
    • 2005
  • Fuzzy logic, neural network, fuzzy-neural network play an important as the key technology of linguistic modeling for intelligent control and decision making in complex systems. The fuzzy-neural network (FNN) learning represents one of the most effective algorithms to build such linguistic models. This paper proposes particle swarm optimization algorithm based optimal learning fuzzy-neural network (PSOA-FNN). The proposed learning scheme is the fuzzy-neural network structure which can handle linguistic knowledge as tuning membership function of fuzzy logic by particle swarm optimization algorithm. The learning algorithm of the PSOA-FNN is composed of two phases. The first phase is to find the initial membership functions of the fuzzy neural network model. In the second phase, particle swarm optimization algorithm is used for tuning of membership functions of the proposed model.

  • PDF

A on-line learning algorithm for recurrent neural networks using variational method (변분법을 이용한 재귀신경망의 온라인 학습)

  • Oh, Oh, Won-Geun;Suh, Suh, Byung-Suhl
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.2 no.1
    • /
    • pp.21-25
    • /
    • 1996
  • In this paper we suggest a general purpose RNN training algorithm which is derived on the optimal control concepts and variational methods. First, learning is regared as an optimal control problem, then using the variational methods we obtain optimal weights which are given by a two-point boundary-value problem. Finally, the modified gradient descent algorithm is applied to RNN for on-line training. This algorithm is intended to be used on learning complex dynamic mappings between time varing I/O data. It is useful for nonlinear control, identification, and signal processing application of RNN because its storage requirement is not high and on-line learning is possible. Simulation results for a nonlinear plant identification are illustrated.

  • PDF

Performance Evaluation of Reinforcement Learning Algorithm for Control of Smart TMD (스마트 TMD 제어를 위한 강화학습 알고리즘 성능 검토)

  • Kang, Joo-Won;Kim, Hyun-Su
    • Journal of Korean Association for Spatial Structures
    • /
    • v.21 no.2
    • /
    • pp.41-48
    • /
    • 2021
  • A smart tuned mass damper (TMD) is widely studied for seismic response reduction of various structures. Control algorithm is the most important factor for control performance of a smart TMD. This study used a Deep Deterministic Policy Gradient (DDPG) among reinforcement learning techniques to develop a control algorithm for a smart TMD. A magnetorheological (MR) damper was used to make the smart TMD. A single mass model with the smart TMD was employed to make a reinforcement learning environment. Time history analysis simulations of the example structure subject to artificial seismic load were performed in the reinforcement learning process. Critic of policy network and actor of value network for DDPG agent were constructed. The action of DDPG agent was selected as the command voltage sent to the MR damper. Reward for the DDPG action was calculated by using displacement and velocity responses of the main mass. Groundhook control algorithm was used as a comparative control algorithm. After 10,000 episode training of the DDPG agent model with proper hyper-parameters, the semi-active control algorithm for control of seismic responses of the example structure with the smart TMD was developed. The simulation results presented that the developed DDPG model can provide effective control algorithms for smart TMD for reduction of seismic responses.

Optimal Control of Induction Motor Using Immune Algorithm Based Fuzzy Neural Network

  • Kim, Dong-Hwa;Cho, Jae-Hoon
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2004.08a
    • /
    • pp.1296-1301
    • /
    • 2004
  • Fuzzy logic, neural network, fuzzy-neural network play an important as the key technology of linguistic modeling for intelligent control and decision making in complex systems. The fuzzy -neural network (FNN) learning represents one of the most effective algorithms to build such linguistic models. This paper proposes learning approach of fuzzy-neural network by immune algorithm. The proposed learning model is presented in an immune based fuzzy-neural network (FNN) form which can handle linguistic knowledge by immune algorithm. The learning algorithm of an immune based FNN is composed of two phases. The first phase used to find the initial membership functions of the fuzzy neural network model. In the second phase, a new immune algorithm based optimization is proposed for tuning of membership functions and structure of the proposed model.

  • PDF

Reward Design of Reinforcement Learning for Development of Smart Control Algorithm (스마트 제어알고리즘 개발을 위한 강화학습 리워드 설계)

  • Kim, Hyun-Su;Yoon, Ki-Yong
    • Journal of Korean Association for Spatial Structures
    • /
    • v.22 no.2
    • /
    • pp.39-46
    • /
    • 2022
  • Recently, machine learning is widely used to solve optimization problems in various engineering fields. In this study, machine learning is applied to development of a control algorithm for a smart control device for reduction of seismic responses. For this purpose, Deep Q-network (DQN) out of reinforcement learning algorithms was employed to develop control algorithm. A single degree of freedom (SDOF) structure with a smart tuned mass damper (TMD) was used as an example structure. A smart TMD system was composed of MR (magnetorheological) damper instead of passive damper. Reward design of reinforcement learning mainly affects the control performance of the smart TMD. Various hyper-parameters were investigated to optimize the control performance of DQN-based control algorithm. Usually, decrease of the time step for numerical simulation is desirable to increase the accuracy of simulation results. However, the numerical simulation results presented that decrease of the time step for reward calculation might decrease the control performance of DQN-based control algorithm. Therefore, a proper time step for reward calculation should be selected in a DQN training process.