• Title/Summary/Keyword: Q-learning system

Search Result 142, Processing Time 0.028 seconds

The Hidden Object Searching Method for Distributed Autonomous Robotic Systems

  • Yoon, Han-Ul;Lee, Dong-Hoon;Sim, Kwee-Bo
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2005.06a
    • /
    • pp.1044-1047
    • /
    • 2005
  • In this paper, we present the strategy of object search for distributed autonomous robotic systems (DARS). The DARS are the systems that consist of multiple autonomous robotic agents to whom required functions are distributed. For instance, the agents should recognize their surrounding at where they are located and generate some rules to act upon by themselves. In this paper, we introduce the strategy for multiple DARS robots to search a hidden object at the unknown area. First, we present an area-based action making process to determine the direction change of the robots during their maneuvers. Second, we also present Q learning adaptation to enhance the area-based action making process. Third, we introduce the coordinate system to represent a robot's current location. In the end of this paper, we show experimental results using hexagon-based Q learning to find the hidden object.

  • PDF

A Case Study of Flipped Learning application of Basics Cooking Practice Subject using YouTube (유튜브를 활용한 기초조리실습과목의 플립드러닝 적용사례 연구)

  • Shin, Seoung-Hoon;Lee, Kyung-Soo
    • The Journal of the Korea Contents Association
    • /
    • v.21 no.5
    • /
    • pp.488-498
    • /
    • 2021
  • This study applied Flipped Learning teaching and learning method to Basics Cooking Practice Subject using YouTube. The purpose of this study is to investigate whether the curriculum is properly progressing by grasping the effects of before and after learning and analyzing learners' subjectivity through the learning process. The investigation period was conducted from August 01, 2020 to September 10, 2020. According to the research design of Q Methodology, it was divided into five stages: Q sample selection, P sample selection, Q sorting, coding and recruiting, conclusion and discussion. As a result of the analysis, the first type (N=5): Prior Learning effect, the second type (N=7): Simulation practice effect, and the third type (N=3): self-efficacy effect. As a result, by applying the flipped learning teaching method of the Basics Cooking Practice Subject using YouTube, positive effects such as inducing interest in the class and increasing confidence were found in active learners, but some learners lacked understanding of the system of the class operation method. However, the lack of number of training sessions compared to other subjects is considered to be a solution to be solved later.

Reinforcement Learning-Based Adaptive Traffic Signal Control considering Vehicles and Pedestrians in Intersection (차량과 보행자를 고려한 강화학습 기반 적응형 교차로 신호제어 연구)

  • Jong-Min Kim;Sun-Yong Kim
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.19 no.1
    • /
    • pp.143-148
    • /
    • 2024
  • Traffic congestion has caused issues in various forms such as the environment and economy. Recently, an intelligent transport system (ITS) using artificial intelligence (AI) has been focused so as to alleviate the traffic congestion problem. In this paper, we propose a reinforcement learning-based traffic signal control algorithm that can smooth the flow of traffic while reducing discomfort levels of drivers and pedestrians. By applying the proposed algorithm, it was confirmed that the discomfort levels of drivers and pedestrians can be significantly reduced compared to the existing fixed signal control system, and that the performance gap increases as the number of roads at the intersection increases.

DEEP LEARNING APPROACH FOR SOLVING A QUADRATIC MATRIX EQUATION

  • Kim, Garam;Kim, Hyun-Min
    • East Asian mathematical journal
    • /
    • v.38 no.1
    • /
    • pp.95-105
    • /
    • 2022
  • In this paper, we consider a quadratic matrix equation Q(X) = AX2 + BX + C = 0 where A, B, C ∈ ℝn×n. A new approach is proposed to find solutions of Q(X), using the novel structure of the information processing system. We also present some numerical experimetns with Artificial Neural Network.

Robust tuning of quadratic criterion-based iterative learning control for linear batch system

  • Kim, Won-Cheol;Lee, Kwang-Soon
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 1996.10a
    • /
    • pp.303-306
    • /
    • 1996
  • We propose a robust tuning method of the quadratic criterion based iterative learning control(Q-ILC) algorithm for discrete-time linear batch system. First, we establish the frequency domain representation for batch systems. Next, a robust convergence condition is derived in the frequency domain. Based on this condition, we propose to optimize the weighting matrices such that the upper bound of the robustness measure is minimized. Through numerical simulation, it is shown that the designed learning filter restores robustness under significant model uncertainty.

  • PDF

Reinforcement Learning-Based Intelligent Decision-Making for Communication Parameters

  • Xie, Xia.;Dou, Zheng;Zhang, Yabin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.9
    • /
    • pp.2942-2960
    • /
    • 2022
  • The core of cognitive radio is the problem concerning intelligent decision-making for communication parameters, the objective of which is to find the most appropriate parameter configuration to optimize transmission performance. The current algorithms have the disadvantages of high dependence on prior knowledge, large amount of calculation, and high complexity. We propose a new decision-making model by making full use of the interactivity of reinforcement learning (RL) and applying the Q-learning algorithm. By simplifying the decision-making process, we avoid large-scale RL, reduce complexity and improve timeliness. The proposed model is able to find the optimal waveform parameter configuration for the communication system in complex channels without prior knowledge. Moreover, this model is more flexible than previous decision-making models. The simulation results demonstrate the effectiveness of our model. The model not only exhibits better decision-making performance in the AWGN channels than the traditional method, but also make reasonable decisions in the fading channels.

An Effective Adaptive Dialogue Strategy Using Reinforcement Loaming (강화 학습법을 이용한 효과적인 적응형 대화 전략)

  • Kim, Won-Il;Ko, Young-Joong;Seo, Jung-Yun
    • Journal of KIISE:Software and Applications
    • /
    • v.35 no.1
    • /
    • pp.33-40
    • /
    • 2008
  • In this paper, we propose a method to enhance adaptability in a dialogue system using the reinforcement learning that reduces response errors by trials and error-search similar to a human dialogue process. The adaptive dialogue strategy means that the dialogue system improves users' satisfaction and dialogue efficiency by loaming users' dialogue styles. To apply the reinforcement learning to the dialogue system, we use a main-dialogue span and sub-dialogue spans as the mathematic application units, and evaluate system usability by using features; success or failure, completion time, and error rate in sub-dialogue and the satisfaction in main-dialogue. In addition, we classify users' groups into beginners and experts to increase users' convenience in training steps. Then, we apply reinforcement learning policies according to users' groups. In the experiments, we evaluated the performance of the proposed method on the individual reinforcement learning policy and group's reinforcement learning policy.

Smart Target Detection System Using Artificial Intelligence (인공지능을 이용한 스마트 표적탐지 시스템)

  • Lee, Sung-nam
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.538-540
    • /
    • 2021
  • In this paper, we proposed a smart target detection system that detects and recognizes a designated target to provide relative motion information when performing a target detection mission of a drone. The proposed system focused on developing an algorithm that can secure adequate accuracy (i.e. mAP, IoU) and high real-time at the same time. The proposed system showed an accuracy of close to 1.0 after 100k learning of the Google Inception V2 deep learning model, and the inference speed was about 60-80[Hz] when using a high-performance laptop based on the real-time performance Nvidia GTX 2070 Max-Q. The proposed smart target detection system will be operated like a drone and will be helpful in successfully performing surveillance and reconnaissance missions by automatically recognizing the target using computer image processing and following the target.

  • PDF

Design of Emotional Learning Controllers for AC Voltage and Circulating Current of Wind-Farm-Side Modular Multilevel Converters

  • Li, Keli;Liao, Yong;Liu, Ren;Zhang, Jimiao
    • Journal of Power Electronics
    • /
    • v.16 no.6
    • /
    • pp.2294-2305
    • /
    • 2016
  • The introduction of a high-voltage direct-current (HVDC) system based on a modular multilevel converter (MMC) for wind farm integration has stimulated studies on methods to control this type of converter. This research article focuses on the control of the AC voltage and circulating current for a wind-farm-side MMC (WFS-MMC). After theoretical analysis, emotional learning (EL) controllers are proposed for the controls. The EL controllers are derived from the learning mechanisms of the amygdala and orbitofrontal cortex which make the WFS-MMC insensitive to variance in system parameters, power change, and fault in the grid. The d-axis and q-axis currents are respectively considered for the d-axis and q-axis voltage controls to improve the performance of AC voltage control. The practicability of the proposed control is verified under various conditions with a point-to-point MMC-HVDC system. Simulation results show that the proposed method is superior to the traditional proportional-integral controller.

Research of Foresight Knowledge by CMAC based Q-learning in Inhomogeneous Multi-Agent System

  • Hoshino, Yukinobu;Sakakura, Akira;Kamei, Katsuari
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2003.09a
    • /
    • pp.280-283
    • /
    • 2003
  • A purpose of our research is an acquisition of cooperative behaviors in inhomogeneous multi-agent system. In this research, we used the fire panic problem as an experiment environment. In Fire panic problem a fire exists in the environment, and follows in each steps of agent's behavior, and this fire spreads within the constant law. The purpose of the agent is to reach the goal established without touching the fire, which exists in the environment. The fire heat up by a few steps, which exists in the environment. The fire has unsureness to the agent. The agent has to avoid a fire, which is spreading in environment. The acquisition of the behavior to reach it to the goal is required. In this paper, we observe how agents escape from the fire cooperating with other agents. For this problem, we propose a unique CMAC based Q-learning system for inhomogeneous multi-agent system.

  • PDF