• 제목/요약/키워드: Control of learning

검색결과 3,767건 처리시간 0.029초

CMAC (Cerebellar Model Arithmetic Controller)

  • Hwang, Heon;Choi, Dong-Y.
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 1989년도 한국자동제어학술회의논문집; Seoul, Korea; 27-28 Oct. 1989
    • /
    • pp.675-681
    • /
    • 1989
  • As an adaptive control function generator, the CMAC (Cerebellar Model Arithmetic or Articulated Controller) based learning control has drawn a great attention to realize a rather robust real-time manipulator control under the various uncertainties. There remain, however, inherent problems to be solved in the CMAC application to robot motion control or perception of sensory information. To apply the CMAC to the various unmodeled or modeled systems more efficiently, It is necessary to analyze the effects of the CMAC control parameters an the trained net. Although the CMAC control parameters such as size of the quantizing block, learning gain, input offset, and ranges of input variables play a key role in the learning performance and system memory requirement, these have not been fully investigated yet. These parameters should be determined, of course, considering the shape of the desired function to be trained and learning algorithms applied. In this paper, the interrelation of these parameters with learning performance is investigated under the basic learning schemes presented by authors. Since an analytic approach only seems to be very difficult and even impossible for this purpose, various simulations have been performed with prespecified functions and their results were analyzed. A general step following design guide was set up according to the various simulation results.

  • PDF

동적 환경에서 강화학습을 이용한 다중이동로봇의 제어 (Reinforcement learning for multi mobile robot control in the dynamic environments)

  • 김도윤;정명진
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 1996년도 한국자동제어학술회의논문집(국내학술편); 포항공과대학교, 포항; 24-26 Oct. 1996
    • /
    • pp.944-947
    • /
    • 1996
  • Realization of autonomous agents that organize their own internal structure in order to behave adequately with respect to their goals and the world is the ultimate goal of AI and Robotics. Reinforcement learning gas recently been receiving increased attention as a method for robot learning with little or no a priori knowledge and higher capability of reactive and adaptive behaviors. In this paper, we present a method of reinforcement learning by which a multi robots learn to move to goal. The results of computer simulations are given.

  • PDF

스마트 TMD 제어를 위한 강화학습 알고리즘 성능 검토 (Performance Evaluation of Reinforcement Learning Algorithm for Control of Smart TMD)

  • 강주원;김현수
    • 한국공간구조학회논문집
    • /
    • 제21권2호
    • /
    • pp.41-48
    • /
    • 2021
  • A smart tuned mass damper (TMD) is widely studied for seismic response reduction of various structures. Control algorithm is the most important factor for control performance of a smart TMD. This study used a Deep Deterministic Policy Gradient (DDPG) among reinforcement learning techniques to develop a control algorithm for a smart TMD. A magnetorheological (MR) damper was used to make the smart TMD. A single mass model with the smart TMD was employed to make a reinforcement learning environment. Time history analysis simulations of the example structure subject to artificial seismic load were performed in the reinforcement learning process. Critic of policy network and actor of value network for DDPG agent were constructed. The action of DDPG agent was selected as the command voltage sent to the MR damper. Reward for the DDPG action was calculated by using displacement and velocity responses of the main mass. Groundhook control algorithm was used as a comparative control algorithm. After 10,000 episode training of the DDPG agent model with proper hyper-parameters, the semi-active control algorithm for control of seismic responses of the example structure with the smart TMD was developed. The simulation results presented that the developed DDPG model can provide effective control algorithms for smart TMD for reduction of seismic responses.

기본간호 실습교육에 있어서 비디오녹화학습의 효과 -배변술을 중심으로- (Effectiveness of Video-Record Method on Fundamental Nursing Skill Education - Focused on Enama -)

  • 강규숙
    • 기본간호학회지
    • /
    • 제3권2호
    • /
    • pp.273-283
    • /
    • 1996
  • Effectiveness of the video-record learning method in teaching bowel elimination nursing skill was investigated using an experimental research methodology. Data was collected from 63 female students attending Fundamental Nursing class from a nursing college in Seoul. The subjects were randomly assigned to two groups, one is the experimental group of 29 and the other the control group of 34. The independent variable was video-record learning method and the dependent variable were the degree of knowledge achivement, nursing skill achivement, competence on practicing elimination skill, and satisfaction about the learning method. The hypotheses of the study were as following. 1) There will be significant difference between the experimental group and the control group in dependent variables. 2) There will be significant positive correlations between nursing skill achievement and other three dependent variables-interest in nursing, adaptation in nursing, and preference of nursing job. Data was analyzed using descriptive statistics, chi-square test, t-test, and Pearson's correlation coefficient with SPSS $PC^+$ program. Findings of the study are : 1) There was no significant difference between the experimental group and the control group in knowledge achievement using P<.05. 2) There was significant difference between the experimental group and the control group in nursing skill achievement using P<.05. 3) There was no significant difference between the experimental group and the control group in competence on practicing elimination skill using P<.05. 4) There was no significant difference between the experimental group and the control group in satisfaction about learning method using P<.05. 5) There was positive correlation between nursing skill achievement and the other variables but no significant difference was shown. 6) This study suggests that video-record learning method is an effective learning method for achiving basic nursing skills but is not effective in other areas such as knowledge achivement, competence in performing nursing practice, and satis-faction about the learning method. Further study with more developed research design and statistical analysis should be done to investigate the effectivenes of video-record learning method in learning basic nursing skill more accurately.

  • PDF

Application Study of Reinforcement Learning Control for Building HVAC System

  • Cho, Sung-Hwan
    • International Journal of Air-Conditioning and Refrigeration
    • /
    • 제14권4호
    • /
    • pp.138-146
    • /
    • 2006
  • Recently, a technology based on the proportional integral (PI) control have grown rapidly owing to the needs for the robust capacity of the controllers from industrial building sectors. However, PI controller generally requires tuning of gains for optimal control when the outside weather condition changes. The present study presents the possibility of reinforcement learning (RL) control algorithm with PI controller adapted in the HVAC system. The optimal design criteria of RL controller was proposed in the environment chamber experiment and a theoretical analysis was also conducted using TRNSYS program.

FUZZY SLIDING MODE ITERATIVE LEARNING CONTROL Of A MANIPULATOR

  • Park, Jae-Sam
    • 대한전자공학회:학술대회논문집
    • /
    • 대한전자공학회 2002년도 ITC-CSCC -3
    • /
    • pp.1483-1486
    • /
    • 2002
  • In this paper, a new scheme of iterative loaming control of a robot manipulator is presented. The proposed method uses a fuzzy sliding mode controller(FSMC), which is designed based on the similarity between the fuzzy logic control(FLC) and the sliding mode control(SMC), for the feedback. With this, the proposed method makes possible fDr fast iteration and has advantages that no linear approximation is used for the derivation of the learning law or in the stability proof Full proof of the convergence of the fuzzy sliding base learning scheme Is given.

  • PDF

ON THE STRUCTURE AND LEARNING OF NEURAL-NETWORK-BASED FUZZY LOGIC CONTROL SYSTEMS

  • C.T. Lin;Lee, C.S. George
    • 한국지능시스템학회:학술대회논문집
    • /
    • 한국퍼지및지능시스템학회 1993년도 Fifth International Fuzzy Systems Association World Congress 93
    • /
    • pp.993-996
    • /
    • 1993
  • This paper addresses the structure and its associated learning algorithms of a feedforward multi-layered connectionist network, which has distributed learning abilities, for realizing the basic elements and functions of a traditional fuzzy logic controller. The proposed neural-network-based fuzzy logic control system (NN-FLCS) can be contrasted with the traditional fuzzy logic control system in their network structure and learning ability. An on-line supervised structure/parameter learning algorithm dynamic learning algorithm can find proper fuzzy logic rules, membership functions, and the size of output fuzzy partitions simultaneously. Next, a Reinforcement Neural-Network-Based Fuzzy Logic Control System (RNN-FLCS) is proposed which consists of two closely integrated Neural-Network-Based Fuzzy Logic Controllers (NN-FLCS) for solving various reinforcement learning problems in fuzzy logic systems. One NN-FLC functions as a fuzzy predictor and the other as a fuzzy controller. As ociated with the proposed RNN-FLCS is the reinforcement structure/parameter learning algorithm which dynamically determines the proper network size, connections, and parameters of the RNN-FLCS through an external reinforcement signal. Furthermore, learning can proceed even in the period without any external reinforcement feedback.

  • PDF

실시간 적응 학습 제어를 위한 진화연산(I) (Evolutionary Computation for the Real-Time Adaptive Learning Control(I))

  • 장성욱;이진걸
    • 대한기계학회:학술대회논문집
    • /
    • 대한기계학회 2001년도 춘계학술대회논문집B
    • /
    • pp.724-729
    • /
    • 2001
  • This paper discusses the composition of the theory of reinforcement learning, which is applied in real-time learning, and evolutionary strategy, which proves its the superiority in the finding of the optimal solution at the off-line learning method. The individuals are reduced in order to learn the evolutionary strategy in real-time, and new method that guarantee the convergence of evolutionary mutations are proposed. It possible to control the control object varied as time changes. As the state value of the control object is generated, applied evolutionary strategy each sampling time because the learning process of an estimation, selection, mutation in real-time. These algorithms can be applied, the people who do not have knowledge about the technical tuning of dynamic systems could design the controller or problems in which the characteristics of the system dynamics are slightly varied as time changes. In the future, studies are needed on the proof of the theory through experiments and the characteristic considerations of the robustness against the outside disturbances.

  • PDF

Robustness of 2nd-order Iterative Learning Control for a Class of Discrete-Time Dynamic Systems

  • 김용태
    • 한국지능시스템학회논문지
    • /
    • 제14권3호
    • /
    • pp.363-368
    • /
    • 2004
  • In this paper, the robustness property of 2nd-order iterative learning control(ILC) method for a class of linear and nonlinear discrete-time dynamic systems is studied. 2nd-order ILC method has the PD-type learning algorithm based on both time-domain performance and iteration-domain performance. It is proved that the 2nd-order ILC method has robustness in the presence of state disturbances, measurement noise and initial state error. In the absence of state disturbances, measurement noise and initialization error, the convergence of the 2nd-order ILC algorithm is guaranteed. A numerical example is given to show the robustness and convergence property according to the learning parameters.

Linear Decentralized Learning Control for the Multiple Dynamic Subsystems

  • Lee, Soo-Cheol
    • 한국산업정보학회논문지
    • /
    • 제1권1호
    • /
    • pp.153-176
    • /
    • 1996
  • The new field of learning control devleops controllers that learn to improve their performance at executing a given task, based on experience performing this task. the simplest forms of learning control are based on the same concepts as integral control, but operating in the domain of the repetitions of the task. This paper studies the use of such controllers ina decentralized system, such as a robot with the controller for each link acting independently. The basic result of the paper is to show that stability of the learning controllers for all subsystems when the coupling between subsystems is turned off, assures stability of the decentralized learning in the coupled system, provided that the sample time in the digital learning controller is sufficiently short.

  • PDF