• 제목/요약/키워드: Learning Control Algorithm

검색결과 944건 처리시간 0.033초

빠른 수렴성을 갖는 로보트 학습제어 (Robot learning control with fast convergence)

  • 양원영;홍호선
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 1988년도 한국자동제어학술회의논문집(국내학술편); 한국전력공사연수원, 서울; 21-22 Oct. 1988
    • /
    • pp.67-71
    • /
    • 1988
  • We present an algorithm that uses trajectory following errors to improve a feedforward command to a robot in the iterative manner. It has been shown that when the manipulator handles an unknown object, the P-type learning algorithm can make the trajectory converge to a desired path and also that the proposed learning control algorithm performs better than the other type learning control algorithm. A numerical simulation of a three degree of freedom manipulator such as PUMA-560 ROBOT has been performed to illustrate the effectiveness of the proposed learning algorithm.

  • PDF

Evolutionary Learning-Rate Selection for BPNN with Window Control Scheme

  • Hoon, Jung-Sung
    • 한국지능시스템학회:학술대회논문집
    • /
    • 한국퍼지및지능시스템학회 1997년도 추계학술대회 학술발표 논문집
    • /
    • pp.301-308
    • /
    • 1997
  • The learning speed of the neural networks, the most important factor in applying to real problems, greatly depends on the learning rate of the networks, Three approaches-empirical, deterministic, and stochastic ones-have been proposed to date. We proposed a new learning-rate selection algorithm using an evolutionary programming search scheme. Even though the performance of our method showed better than those of the other methods, it was found that taking much time for selecting evolutionary learning rates made the performance of our method degrade. This was caused by using static intervals (called static windows) in order to update learning rates. Out algorithm with static windows updated the learning rates showed good performance or didn't update the learning rates even though previously updated learning rates shoved bad performance. This paper introduce a window control scheme to avoid such problems. With the window control scheme, our algorithm try to update the learning ra es only when the learning performance is continuously bad during a specified interval. If previously selected learning rates show good performance, new algorithm will not update the learning rates. This diminish the updating time of learning rates greatly. As a result, our algorithm with the window control scheme show better performance than that with static windows. In this paper, we will describe the previous and new algorithm and experimental results.

  • PDF

직접 구동형 매니퓰레이터를 위한 학습 제어기의 실시간 구현에 관한 연구 (A Study on Implementation of a Real Time Learning Controller for Direct Drive Manipulator)

  • 전종욱;안현식;임미섭;김권호;김광배;이쾌희
    • 대한전기학회:학술대회논문집
    • /
    • 대한전기학회 1993년도 하계학술대회 논문집 A
    • /
    • pp.369-372
    • /
    • 1993
  • In this thesis, we consider an iterative learning controller to control the continuous trajectory of 2 links direct drive robot manipulator and process computer simulation and real-time experiment. To improve control performance, we adapt an iterative learning control algorithm, drive a sufficient condition for convergence from which is drived extended conventional control algorithm and get better performance by extended learning control algorithm than that by conventional algorithm from simulation results. Also, experimental results show that better performance is taken by extended learning algorithm.

  • PDF

선형 회분식 공정을 위한 이차 성능 지수에 의한 모델 기반 반복 학습 제어 (Model-based iterative learning control with quadratic criterion for linear batch processes)

  • 이광순;김원철;이재형
    • 제어로봇시스템학회논문지
    • /
    • 제2권3호
    • /
    • pp.148-157
    • /
    • 1996
  • Availability of input trajectories corresponding to desired output trajectories is often important in designing control systems for batch and other transient processes. In this paper, we propose a predictive control-type model-based iterative learning algorithm which is applicable to finding the nominal input trajectories of a linear time-invariant batch process. Unlike the other existing learning control algorithms, the proposed algorithm can be applied to nonsquare systems and has an ability to adjust noise sensitivity as well as convergence rate. A simple model identification technique with which performance of the proposed learning algorithm can be significantly enhanced is also proposed. Performance of the proposed learning algorithm is demonstrated through numerical simulations.

  • PDF

DOA 기반 학습률 조절을 이용한 다채널 음성개선 알고리즘 (Multi-Channel Speech Enhancement Algorithm Using DOA-based Learning Rate Control)

  • 김수환;이영재;김영일;정상배
    • 말소리와 음성과학
    • /
    • 제3권3호
    • /
    • pp.91-98
    • /
    • 2011
  • In this paper, a multi-channel speech enhancement method using the linearly constrained minimum variance (LCMV) algorithm and a variable learning rate control is proposed. To control the learning rate for adaptive filters of the LCMV algorithm, the direction of arrival (DOA) is measured for each short-time input signal and the likelihood function of the target speech presence is estimated to control the filter learning rate. Using the likelihood measure, the learning rate is increased during the pure noise interval and decreased during the target speech interval. To optimize the parameter of the mapping function between the likelihood value and the corresponding learning rate, an exhaustive search is performed using the Bark's scale distortion (BSD) as the performance index. Experimental results show that the proposed algorithm outperforms the conventional LCMV with fixed learning rate in the BSD by around 1.5 dB.

  • PDF

Optimal Learning of Fuzzy Neural Network Using Particle Swarm Optimization Algorithm

  • Kim, Dong-Hwa;Cho, Jae-Hoon
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2005년도 ICCAS
    • /
    • pp.421-426
    • /
    • 2005
  • Fuzzy logic, neural network, fuzzy-neural network play an important as the key technology of linguistic modeling for intelligent control and decision making in complex systems. The fuzzy-neural network (FNN) learning represents one of the most effective algorithms to build such linguistic models. This paper proposes particle swarm optimization algorithm based optimal learning fuzzy-neural network (PSOA-FNN). The proposed learning scheme is the fuzzy-neural network structure which can handle linguistic knowledge as tuning membership function of fuzzy logic by particle swarm optimization algorithm. The learning algorithm of the PSOA-FNN is composed of two phases. The first phase is to find the initial membership functions of the fuzzy neural network model. In the second phase, particle swarm optimization algorithm is used for tuning of membership functions of the proposed model.

  • PDF

변분법을 이용한 재귀신경망의 온라인 학습 (A on-line learning algorithm for recurrent neural networks using variational method)

  • 오원근;서병설
    • 제어로봇시스템학회논문지
    • /
    • 제2권1호
    • /
    • pp.21-25
    • /
    • 1996
  • In this paper we suggest a general purpose RNN training algorithm which is derived on the optimal control concepts and variational methods. First, learning is regared as an optimal control problem, then using the variational methods we obtain optimal weights which are given by a two-point boundary-value problem. Finally, the modified gradient descent algorithm is applied to RNN for on-line training. This algorithm is intended to be used on learning complex dynamic mappings between time varing I/O data. It is useful for nonlinear control, identification, and signal processing application of RNN because its storage requirement is not high and on-line learning is possible. Simulation results for a nonlinear plant identification are illustrated.

  • PDF

스마트 TMD 제어를 위한 강화학습 알고리즘 성능 검토 (Performance Evaluation of Reinforcement Learning Algorithm for Control of Smart TMD)

  • 강주원;김현수
    • 한국공간구조학회논문집
    • /
    • 제21권2호
    • /
    • pp.41-48
    • /
    • 2021
  • A smart tuned mass damper (TMD) is widely studied for seismic response reduction of various structures. Control algorithm is the most important factor for control performance of a smart TMD. This study used a Deep Deterministic Policy Gradient (DDPG) among reinforcement learning techniques to develop a control algorithm for a smart TMD. A magnetorheological (MR) damper was used to make the smart TMD. A single mass model with the smart TMD was employed to make a reinforcement learning environment. Time history analysis simulations of the example structure subject to artificial seismic load were performed in the reinforcement learning process. Critic of policy network and actor of value network for DDPG agent were constructed. The action of DDPG agent was selected as the command voltage sent to the MR damper. Reward for the DDPG action was calculated by using displacement and velocity responses of the main mass. Groundhook control algorithm was used as a comparative control algorithm. After 10,000 episode training of the DDPG agent model with proper hyper-parameters, the semi-active control algorithm for control of seismic responses of the example structure with the smart TMD was developed. The simulation results presented that the developed DDPG model can provide effective control algorithms for smart TMD for reduction of seismic responses.

Optimal Control of Induction Motor Using Immune Algorithm Based Fuzzy Neural Network

  • Kim, Dong-Hwa;Cho, Jae-Hoon
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2004년도 ICCAS
    • /
    • pp.1296-1301
    • /
    • 2004
  • Fuzzy logic, neural network, fuzzy-neural network play an important as the key technology of linguistic modeling for intelligent control and decision making in complex systems. The fuzzy -neural network (FNN) learning represents one of the most effective algorithms to build such linguistic models. This paper proposes learning approach of fuzzy-neural network by immune algorithm. The proposed learning model is presented in an immune based fuzzy-neural network (FNN) form which can handle linguistic knowledge by immune algorithm. The learning algorithm of an immune based FNN is composed of two phases. The first phase used to find the initial membership functions of the fuzzy neural network model. In the second phase, a new immune algorithm based optimization is proposed for tuning of membership functions and structure of the proposed model.

  • PDF

스마트 제어알고리즘 개발을 위한 강화학습 리워드 설계 (Reward Design of Reinforcement Learning for Development of Smart Control Algorithm)

  • 김현수;윤기용
    • 한국공간구조학회논문집
    • /
    • 제22권2호
    • /
    • pp.39-46
    • /
    • 2022
  • Recently, machine learning is widely used to solve optimization problems in various engineering fields. In this study, machine learning is applied to development of a control algorithm for a smart control device for reduction of seismic responses. For this purpose, Deep Q-network (DQN) out of reinforcement learning algorithms was employed to develop control algorithm. A single degree of freedom (SDOF) structure with a smart tuned mass damper (TMD) was used as an example structure. A smart TMD system was composed of MR (magnetorheological) damper instead of passive damper. Reward design of reinforcement learning mainly affects the control performance of the smart TMD. Various hyper-parameters were investigated to optimize the control performance of DQN-based control algorithm. Usually, decrease of the time step for numerical simulation is desirable to increase the accuracy of simulation results. However, the numerical simulation results presented that decrease of the time step for reward calculation might decrease the control performance of DQN-based control algorithm. Therefore, a proper time step for reward calculation should be selected in a DQN training process.