• Title/Summary/Keyword: Multi-learning System

Search Result 620, Processing Time 0.028 seconds

A Performance Improvement Technique for Nash Q-learning using Macro-Actions (매크로 행동을 이용한 내시 Q-학습의 성능 향상 기법)

  • Sung, Yun-Sik;Cho, Kyun-Geun;Um, Ky-Hyun
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.3
    • /
    • pp.353-363
    • /
    • 2008
  • A multi-agent system has a longer learning period and larger state-spaces than a sin91e agent system. In this paper, we suggest a new method to reduce the learning time of Nash Q-learning in a multi-agent environment. We apply Macro-actions to Nash Q-learning to improve the teaming speed. In the Nash Q-teaming scheme, when agents select actions, rewards are accumulated like Macro-actions. In the experiments, we compare Nash Q-learning using Macro-actions with general Nash Q-learning. First, we observed how many times the agents achieve their goals. The results of this experiment show that agents using Nash Q-learning and 4 Macro-actions have 9.46% better performance than Nash Q-learning using only 4 primitive actions. Second, when agents use Macro-actions, Q-values are accumulated 2.6 times more. Finally, agents using Macro-actions select less actions about 44%. As a result, agents select fewer actions and Macro-actions improve the Q-value's update. It the agents' learning speeds improve.

  • PDF

Online Education System for Work Based Learning Dual System (일-학습 병행을 위한 온라인 교육 시스템)

  • Kwon, Oh-Young
    • Journal of Practical Engineering Education
    • /
    • v.5 no.2
    • /
    • pp.163-168
    • /
    • 2013
  • The vicious cycle of over-education has been made. That is, higher education enrollment rate is high, but university graduate employment rate is low. To eliminate this cycle and relieve youth unemployment and young people to enter the labor market early, dual education and training system is needed. This dual system can support working and learning in parallel. So, worker can get the opportunity pre-employment and post-learning and improve his/her job skills. Recent MOOC (Massive Open On-line Course), a new form of online education system, has emerged. MOOC combines education, entertainment and social networking, and emphasize the interaction between faculty and student and between students. The educational contents of MOOC are available free of charge. Using newly changed online education environments we can effectively provide knowledge and skills. In technology and engineering education hands-on training is necessary. In order to support work based learning dual system for worker to work and learn in parallel, we should build the multi-learning system to combine the online education and campus hands-on practice.

Universal learning network-based fuzzy control

  • Hirasawa, K.;Wu, R.;Ohbayashi, M.
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1995.10a
    • /
    • pp.436-439
    • /
    • 1995
  • In this paper we present a method to construct fuzzy model with multi-dimension input membership function, which can construct fuzzy inference system on one node of the network directly. This method comes from a common framework called Universal Learning Network (ULN). The fuzzy model under the framework of ULN is called Universal Learning Network-based Fuzzy Inference System (ULNFIS), which possesses certain advantages over other networks such as neural network. We also introduce how to imitate a real system with ULN and a control scheme using ULNFIS.

  • PDF

Q-learning for intersection traffic flow Control based on agents

  • Zhou, Xuan;Chong, Kil-To
    • Proceedings of the IEEK Conference
    • /
    • 2009.05a
    • /
    • pp.94-96
    • /
    • 2009
  • In this paper, we present the Q-learning method for adaptive traffic signal control on the basis of multi-agent technology. The structure is composed of sixphase agents and one intersection agent. Wireless communication network provides the possibility of the cooperation of agents. As one kind of reinforcement learning, Q-learning is adopted as the algorithm of the control mechanism, which can acquire optical control strategies from delayed reward; furthermore, we adopt dynamic learning method instead of static method, which is more practical. Simulation result indicates that it is more effective than traditional signal system.

  • PDF

Exercise Recommendation System Using Deep Neural Collaborative Filtering (신경망 협업 필터링을 이용한 운동 추천시스템)

  • Jung, Wooyong;Kyeong, Chanuk;Lee, Seongwoo;Kim, Soo-Hyun;Sun, Young-Ghyu;Kim, Jin-Young
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.22 no.6
    • /
    • pp.173-178
    • /
    • 2022
  • Recently, a recommendation system using deep learning in social network services has been actively studied. However, in the case of a recommendation system using deep learning, the cold start problem and the increased learning time due to the complex computation exist as the disadvantage. In this paper, the user-tailored exercise routine recommendation algorithm is proposed using the user's metadata. Metadata (the user's height, weight, sex, etc.) set as the input of the model is applied to the designed model in the proposed algorithms. The exercise recommendation system model proposed in this paper is designed based on the neural collaborative filtering (NCF) algorithm using multi-layer perceptron and matrix factorization algorithm. The learning proceeds with proposed model by receiving user metadata and exercise information. The model where learning is completed provides recommendation score to the user when a specific exercise is set as the input of the model. As a result of the experiment, the proposed exercise recommendation system model showed 10% improvement in recommended performance and 50% reduction in learning time compared to the existing NCF model.

A Meta-learning Approach for Building Multi-classifier Systems in a GA-based Inductive Learning Environment (유전 알고리즘 기반 귀납적 학습 환경에서 다중 분류기 시스템의 구축을 위한 메타 학습법)

  • Kim, Yeong-Joon;Hong, Chul-Eui
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.19 no.1
    • /
    • pp.35-40
    • /
    • 2015
  • The paper proposes a meta-learning approach for building multi-classifier systems in a GA-based inductive learning environment. In our meta-learning approach, a classifier consists of a general classifier and a meta-classifier. We obtain a meta-classifier from classification results of its general classifier by applying a learning algorithm to them. The role of the meta-classifier is to evaluate the classification result of its general classifier and decide whether to participate into a final decision-making process or not. The classification system draws a decision by combining classification results that are evaluated as correct ones by meta-classifiers. We present empirical results that evaluate the effect of our meta-learning approach on the performance of multi-classifier systems.

Learning Control of Inverted Pendulum Using Neural Networks (신경회로망을 이용한 도립전자의 학습제어)

  • Lee, Jea-Kang;Kim, Il-Hwan
    • Journal of Industrial Technology
    • /
    • v.24 no.A
    • /
    • pp.99-107
    • /
    • 2004
  • This paper considers reinforcement learning control with the self-organizing map. Reinforcement learning uses the observable states of objective system and signals from interaction of the system and the environments as input data. For fast learning in neural network training, it is necessary to reduce learning data. In this paper, we use the self-organizing map to parition the observable states. Partitioning states reduces the number of learning data which is used for training neural networks. And neural dynamic programming design method is used for the controller. For evaluating the designed reinforcement learning controller, an inverted pendulum of the cart system is simulated. The designed controller is composed of serial connection of self-organizing map and two Multi-layer Feed-Forward Neural Networks.

  • PDF

Design of Reinforcement Learning Controller with Self-Organizing Map (자기 조직화 맵을 이용한 강화학습 제어기 설계)

  • 이재강;김일환
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.53 no.5
    • /
    • pp.353-360
    • /
    • 2004
  • This paper considers reinforcement learning control with the self-organizing map. Reinforcement learning uses the observable states of objective system and signals from interaction of the system and environment as input data. For fast learning in neural network training, it is necessary to reduce learning data. In this paper, we use the self-organizing map to partition the observable states. Partitioning states reduces the number of learning data which is used for training neural networks. And neural dynamic programming design method is used for the controller. For evaluating the designed reinforcement learning controller, an inverted pendulum on the cart system is simulated. The designed controller is composed of serial connection of self-organizing map and two Multi-layer Feed-Forward Neural Networks.

A Markov Decision Process (MDP) based Load Balancing Algorithm for Multi-cell Networks with Multi-carriers

  • Yang, Janghoon
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.8 no.10
    • /
    • pp.3394-3408
    • /
    • 2014
  • Conventional mobile state (MS) and base station (BS) association based on average signal strength often results in imbalance of cell load which may require more powerful processor at BSs and degrades the perceived transmission rate of MSs. To deal with this problem, a Markov decision process (MDP) for load balancing in a multi-cell system with multi-carriers is formulated. To solve the problem, exploiting Sarsa algorithm of on-line learning type [12], ${\alpha}$-controllable load balancing algorithm is proposed. It is designed to control tradeoff between the cell load deviation of BSs and the perceived transmission rates of MSs. We also propose an ${\varepsilon}$-differential soft greedy policy for on-line learning which is proven to be asymptotically convergent to the optimal greedy policy under some condition. Simulation results verify that the ${\alpha}$-controllable load balancing algorithm controls the behavior of the algorithm depending on the choice of ${\alpha}$. It is shown to be very efficient in balancing cell loads of BSs with low ${\alpha}$.

A Study on the Prediction Diagnosis System Improvement by Error Terms and Learning Methodologies Application (오차항과 러닝 기법을 활용한 예측진단 시스템 개선 방안 연구)

  • Kim, Myung Joon;Park, Youngho;Kim, Tai Kyoo;Jung, Jae-Seok
    • Journal of Korean Society for Quality Management
    • /
    • v.47 no.4
    • /
    • pp.783-793
    • /
    • 2019
  • Purpose: The purpose of this study is to apply the machine and deep learning methodology on error terms which are continuously auto-generated on the sensors with specific time period and prove the improvement effects of power generator prediction diagnosis system by comparing detection ability. Methods: The SVM(Support Vector Machine) and MLP(Multi Layer Perception) learning procedures were applied for predicting the target values and sequentially producing the error terms for confirming the detection improvement effects of suggested application. For checking the effectiveness of suggested procedures, several detection methodologies such as Cusum and EWMA were used for the comparison. Results: The statistical analysis result shows that without noticing the sequential trivial changes on current diagnosis system, suggested approach based on the error term diagnosis is sensing the changes in the very early stages. Conclusion: Using pattern of error terms as a diagnosis tool for the safety control process with SVM and MLP learning procedure, unusual symptoms could be detected earlier than current prediction system. By combining the suggested error term management methodology with current process seems to be meaningful for sustainable safety condition by early detecting the symptoms.